You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Just my two cents but might it be because these models tend to keep repeating: issue on TRT for llama, and for mixtral or maybe I have just confirmation bias.
Is there any reason why we have an accuracy upper limit for LLAMA2 Tokens per sample but not for GPT-J? It's good to document this reason for users.
The text was updated successfully, but these errors were encountered: