You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
notable pcc drop of 0.9722469167498842 is observed in the google/flan-t5-large model variant,By running decoder alone drop of 0.9857664543253748 is observed.
A detailed analysis of PCC changes with respect to the number of decoder layers reveals the following
Number of decoder layers
pcc
1
1.0
12
0.9999998585756898
18
0.9998312374946592
23
0.9939849829870099
full decoder (24)
0.9857664543253748
From the table, it is evident that the PCC drop is negligible in the earlier decoder layers but becomes notable only at the final decoder block. Despite the small and gradual decrease as the layer count increases, pinpointing the operation responsible for this issue has proven challenging. Further investigation is needed to isolate and address the root cause.
Reproduce
git checkout kkannan/t5_fix_dec2
git submodule update --recursive
cmake --build build -- install_ttforge
# to run whole model
pytest forge/test/models/pytorch/text/t5/test_t5.py::test_t5_generation[no_device-google_flan_t5_large] -svv
# to run decoder alone
pytest forge/test/models/pytorch/text/t5/test_t5_sanity.py -svv
Note : twig the num_decoder_layers to get table results
Observed Behaviour
Tensor mismatch on Framework vs. Forge codegen output 0
Max ATOL Delta: 7.033e+00, atol=5.368640899658203
Max RTOL Delta: 4.264e+00, rtol=0
PCC got=0.9722469167498842, required=0.99999
Description
notable pcc drop of 0.9722469167498842 is observed in the google/flan-t5-large model variant,By running decoder alone drop of 0.9857664543253748 is observed.
A detailed analysis of PCC changes with respect to the number of decoder layers reveals the following
Reproduce
Observed Behaviour
Logs
The text was updated successfully, but these errors were encountered: