[torchbench] hf_GPT2
(large, too) fails to run on bfloat16
dtype.
#6521
Labels
hf_GPT2
(large, too) fails to run on bfloat16
dtype.
#6521
🐛 Bug
After converting the
hf_GPT2
(and its large variation) model tobfloat16
and running it (see command below), it fails with the following error:python xla/benchmarks/experiment_runner.py \ --suite-name torchbench --accelerator cuda \ --xla PJRT --dynamo None --test eval \ --no-resume --print-subprocess \ -k hf_GPT2
To Reproduce
bfloat16
(inference) and AMP (training) precision. #6518Affected Configurations
Environment
cc @miladm
The text was updated successfully, but these errors were encountered: