-
Notifications
You must be signed in to change notification settings - Fork 191
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to run transformer-xl with parallel experts with single gpu? #211
Comments
It is supported using multiple cuda streams. Refer to the class |
多谢回复~ |
可以设成更大的数, 能跑. |
可以了,确实很强大。我的配置上,代码比for-loop快了3倍以上,而且topk基本不影响速度,赞👍🏻 |
@laekov 我把gpu设置成cuda:1的话,就会报错:
|
可能有的数据还在 cuda:0 上? |
seems fast-moe still cannot archive running multi experts in parallel with single gpu card?
The text was updated successfully, but these errors were encountered: