-
Notifications
You must be signed in to change notification settings - Fork 520
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix AMD FP8 Test and use native rowwise quantization in benchmark #2849
Conversation
✅ Deploy Preview for pytorch-fbgemm-docs ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
This pull request was exported from Phabricator. Differential Revision: D59771162 |
This pull request was exported from Phabricator. Differential Revision: D59771162 |
…torch#2849) Summary: Pull Request resolved: pytorch#2849 Fix a minor test issue where triton blockwise quantization was running on AMD despite not being supported. I also switch rowwise quantization in our fp8 benchmarks to the native hip implementation. Reviewed By: jianyuh Differential Revision: D59771162
37ba9ec
to
dd051db
Compare
This pull request was exported from Phabricator. Differential Revision: D59771162 |
…torch#2849) Summary: Pull Request resolved: pytorch#2849 Fix a minor test issue where triton blockwise quantization was running on AMD despite not being supported. I also switch rowwise quantization in our fp8 benchmarks to the native hip implementation. Reviewed By: jianyuh Differential Revision: D59771162
dd051db
to
2cbdfbe
Compare
…torch#2849) Summary: Pull Request resolved: pytorch#2849 Fix a minor test issue where triton blockwise quantization was running on AMD despite not being supported. I also switch rowwise quantization in our fp8 benchmarks to the native hip implementation. Reviewed By: jianyuh Differential Revision: D59771162
This pull request was exported from Phabricator. Differential Revision: D59771162 |
2cbdfbe
to
75ff00a
Compare
This pull request has been merged in 57a5969. |
Summary:
Fix a minor test issue where triton blockwise quantization was running on AMD despite not being supported.
I also switch rowwise quantization in our fp8 benchmarks to the native hip implementation.
Reviewed By: jianyuh
Differential Revision: D59771162