Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Torch 2.0 #169

Open
varshith15 opened this issue Apr 11, 2023 · 3 comments
Open

Torch 2.0 #169

varshith15 opened this issue Apr 11, 2023 · 3 comments

Comments

@varshith15
Copy link

varshith15 commented Apr 11, 2023

Is there a way to leverage torch2.0's compile using tensorrt as a backend directly? without all the current tedious process? https://pytorch.org/docs/stable/dynamo/get-started.html

And any thoughts on torch 2.0 in general? Has anyone tried it out?
I've tried it out for a few of the transformer models, there doesn't seem to be any improvement.
@pommedeterresautee @ayoub-louati

@pommedeterresautee
Copy link
Member

Yes tensorRT is supported out of the box. However, it adds its own overhead and is not always best choice in my tests. Kernl runs on top of PyTorch 2.0. The 2.0 targets mostly for now training (and not inference).

@varshith15
Copy link
Author

@pommedeterresautee any thoughts on Apache tvm?

@pommedeterresautee
Copy link
Member

pommedeterresautee commented Apr 26, 2023

TVM was best for non GPU stuff. Recently they started to support better GPU through cutlass + adding possibility to program at block of threads level (CTAs), but IMO Triton is a better choice for now when Nvidia hw is your target

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants