XRT branch cherry-pick 07/10 (#5293)
* Skip calling as_strided in empty_strided_symint if the input has dynamic dimensions. (#5239)
* Skip calling as_strided in empty_strided_symint.
* only return empty_symint conditionally.
* add a comment
* Add XRT nightly builds (#5261)
* Add XRT nightly builds
* remove space
* Add ToString method for both PjrtData and PjrtShardedData (#5265)
* Add ToString method for both PjrtData and PjrtShardedData
* on cpu same config will become replicated, dont't check actual op sharding type
* fix xrt tostring
* Update Sharded graph HLO dumping (#5266)
* Disable Bazel remote cache for forked PR (#5259)
* disable bazel remote cache if gcloud key is empty
* remove remote cache from setup.py
* experiment with debug msg
* fix flag
* add more logs
* skip remote chache if credential file is empty
* add comment
* add logs
* add check in test and coverage script
* fix condition in coverage test
* advance branch pr
* allow remote cache if gloud file isn't specified explicitly
* remove dummy comment
* Suppress debug symbols in OpenXLA code (#5269)
* [SPMD] Sharding n-d tensor on (n+1)-d Mesh (#5268)
* Make TPU detection more robust (#5271)
* Clean bazel stuff on distutils clean. (#5274)
* Clean bazel stuff on distutils clean
* Fix python formatting
* fix conflict
* Fix the error when export_torch_model is given a non-tensor (#5277)
However the generated StableHLO graph still hardcodes the
non-tensor value. this is not correct, will fix later.
* Dsiable test_simple_model_with_different_input_shape since it is curretnly broken by pytorch (#5282)
* Always do build_ext in python setup.py develop (#5273)
Bazel should figure out that _XLAC.so is current
or not, and trigger rebuild if any cpp files changed.
* Remove or improve several hardcoded TPU test conditions (#5272)
* Remove or improve several hardcoded TPU test conditions
* Fix test condition
* Add `runtime.host_index` (#5283)
* Make it an error if calling sizes() on a dynamic tensor. (#4998)
* Err if calling sizes() on dynamic tensor
* try to set has_symbolic_sizes_strides_
* resolve merge conflict
* enable CONTINUE_ON_ERROR
* fixed the python test test_SizeEq_should_not_compile_for_identical_symints
* fix test_index_types
* set CONTINUE_ON_ERROR to true
* remove some unwanted code.
* add a print
* directly set has_symbolic_sizes_strides_ = true
* make some fixes.
* fix empty_strided_symint
* ran linter
* change error type in the test.
* fix comments
* ran linter
* Fix the error where mark_step does not materalize tensors on SPMD:0 (#5281)
* Fix the error where mark_step does not materalize tensors on SPMD:0
* typo
* fix test_non_tensor_scalar
* Disable torch._dynamo.config.automatic_dynamic_shapes (#5285)
* Set torch._dynamo.config.automatic_dynamic_shapes to False
* Enable DynamoInferenceBasicTest.test_simple_model_with_different_input_shape
* [Traceable Collecive] Hide token for all_gather (#5232)
Summary:
This pull request does the following:
1. It hides token for all_gather.
2. It folds the out-of-place all_gather into the regular all_gather.
3. It fixes an issue with the last all_reduce_in_place PR where it forgot to set the token.
Test Plan:
PJRT_DEVICE=TPU python test/test_mp_all_gather.py
* Lower squeeze.dims (#5286)
* avoid copy proto in PrepareOutputShardingPropagation (#5287)
* Revert "Suppress debug symbols in OpenXLA code (#5269)"
This reverts commit 3967d7ba2d29653b0468567464e8e4e4f399869b.
* Revert "fix conflict"
This reverts commit e91ad3ad412e3251b243626baaea4fcbb3c9b5e1.
---------
Co-authored-by: iefgnoix <isaacwxf23@gmail.com>
Co-authored-by: Will Cromar <wcromar@google.com>
Co-authored-by: Siyuan Liu <lsiyuan@google.com>
Co-authored-by: stgpetrovic <stgpetrovic@gmail.com>
Co-authored-by: Mohit Khatwani <118776932+khatwanimohit@users.noreply.github.com>
Co-authored-by: qihqi <hanq@google.com>
Co-authored-by: Wonjoo Lee <wonjoo@google.com>
Co-authored-by: Jiewen Tan <jwtan@google.com>
Co-authored-by: Baole Ai <baoleai01@gmail.com>