forked from pytorch/xla
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
tpu ci module refactor #7
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Testing flux
test branch
test branch
Test branch
Fix kubernetes provider
…h#5751) * fix squeeze op lowering issue when dim is not in sorted order * remove debug info * remove debug info * refactor BuildSqueezedDimensions
…ytorch#5777) * Move pure dtype conversion functions to `dtype.cpp` * remove comments * better names * fix includes * formatting * consolidate * fix test build * more explicit names * remove extra line
…Module (pytorch#5745) Co-authored-by: Siyuan Liu <lsiyuan@google.coim>
* delete nccl_distributed * remove async_task * remove unique * Remove hashing * more random cleanup * formatting * remove util.cc * Revert "remove unique" This reverts commit ebe4567. * Use upstream Unique
* Make the pjrt gpu allocator configurable * the default value changed from 0.9 to 0.75 * return default GpuAllocatorConfig --------- Co-authored-by: wangang.wa <wangang.wa@alibaba-inc.com>
* [SPMD] move SPMD package to torch_xla/experimental/spmd, introduce shadow xla DTensor API. * support backward compatibility of the old imports * Move spmd out of experimental * Update spmd.md for distributed/spmd
* Transfer data directly to the device (pytorch#5752) * Remove `populate_fn` from `TensorSource` * Make TensorSource an interface * Re-enable pjrt_computation_client_test * server -> device * add comment * fix outbound data metric * formatting * implement byte_strides in TensorSource * more formatting * remove extra deps * add missing deps * Revert "server -> device" This reverts commit 6384516. * Use `at::Tensor`'s layout for byte strides * Downcast at::Tensor if required * formatting * Simplify AtenSource * fix build * formatting * fix typo that makes us ignore input type * Revert "Simplify AtenSource" This reverts commit 4225deb. * Skip hanging test * fix gil deadlock * formatting
* lower full * update test for full op * formatting
…er (pytorch#5770) * Add GKE support and various usability improvements in CheckpointManager * Bug fix for async checkpointing fully sharded state dicts
* Record the lazy tracing time(C++) in metrics * Delete torch_patches/.torch_pin
This reverts commit 4baef3c.
* port sandeep unbounded dynamism change * Enable unbounded dynamism using env var, add more guards for unbounded dynamism code path --------- Co-authored-by: Siyuan Liu <lsiyuan@google.coim>
* Use TSL threadpool * remove multiwait * fix test build * Move threadpool namespace * formatting * fix test build * Use BlockingCounter
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
masked_fill
broadcasting. (Fixmasked_fill
broadcasting. pytorch/xla#5688)__init__.py
(Conditionally set default TPU settings in__init__.py
pytorch/xla#5696)_unsafe_index
. (Add support for_unsafe_index
. pytorch/xla#5707)--pjrt_distributed
flag with--ddp
flag. (Merge--pjrt_distributed
flag with--ddp
flag. pytorch/xla#5732)XlaDeviceToAtenDevice
. (Fix XLA tensor storage device by usingXlaDeviceToAtenDevice
. pytorch/xla#5743)_unsafe_index
implementation. (Remove_unsafe_index
implementation. pytorch/xla#5769)//torch_xla/csrc:tensor
(Refactor type conversion functions out of//torch_xla/csrc:tensor
pytorch/xla#5777)csrc/runtime
(Remove some unused code fromcsrc/runtime
pytorch/xla#5785)