-
Notifications
You must be signed in to change notification settings - Fork 467
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pin update March 2024 #6677
Pin update March 2024 #6677
Conversation
Hit the following error
Need to upgrade bazel version to above 6.0.0 |
b70f43f
to
069f708
Compare
Testing performance with the following cmd, on v4-8 TPU
After pin update
Before pin update
There is a perf regression after the pin update. Update: The perf result above is using debugging build, redo with release build.
Before pin upate
|
Test failed with PT2E test, because the converter patch is commented out now. Move xla pin again after https://github.com/pytorch/xla/blob/master/openxla_patches/quant_dequant_converter.diff is upstreamed |
The following GPU tests hit OOM in CI after pin update
Example error message:
|
b8b4aaa
to
1e48f58
Compare
1e48f58
to
e857c3e
Compare
This reverts commit e857c3e.
…he change in converter
689e2dc
to
e39587c
Compare
cc @will-cromar for some PJRT changes to accommodate the change of PJRT interface in upstream XLA . |
Thanks @sdasgup3 for pointing out we need to generate custom call to |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Co-authored-by: Siyuan Liu <lsiyuan@google.coim>
Update xla pin to HEAD
Summary:
PJRT_Structure_Base
toPJRT_Extension_Base
to accommodate change in XLA.mhlo.uniform_dequantize/quantize
to accommodate the change in HLO->MHLO converter patch.