GPT throughput (v.s. Megatron-LM) #2842
Replies: 3 comments 1 reply
-
Hi @yurishin929 Titans is our temporary model zoo because some complicate parallel strategies may require users to modify the model. We offer some mainstream models for users in Titans. |
Beta Was this translation helpful? Give feedback.
-
Hi @binmakeswell ,
Thanks in advance! |
Beta Was this translation helpful? Give feedback.
-
Hi @Agoniii Using Gemini with |
Beta Was this translation helpful? Give feedback.
-
Hi,
In README, when using GPT-3 model, colossal-ai shows better performance(sec/iter and Thoughput) than Megatron-LM. and Thoughput for GPT-2, Bert too.
so my question is, for these performances, did you guys use Titans?
especially in Colossal-AI for GPT-3 Table, second one that shows 4.99 throuput.
I'm using Gemini and Megatron-LM for GPT2-medium. and Gemini has worse throughput than Megatron-LM(but better memory efficient). Does Titans show better thoughput(and lower sec/iter) than Gemini and Megatron? Thank you.
Beta Was this translation helpful? Give feedback.
All reactions