-
Notifications
You must be signed in to change notification settings - Fork 162
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ImportError: cannot import name '_grouped_size_compiled_for_decode_kernels' from 'flashinfer.decode' #549
Comments
That function was removed recently (because with the new JIT feature, all group size can be compiled with JIT) in mainline, I can add this back for backward compatibility, but it's better to not rely on this function in sglang. We can use some heuristic to control whether to use tensor cores or not. @hnyls2002 @merrymercy WDTY? |
Or you can install our pre-built wheels through pypi, or install from v0.1.6 source at https://github.com/flashinfer-ai/flashinfer/tree/v0.1.6 |
Thank you very much, I will try this method and hope it works. |
@yzh119 Can you provide a utility function inside flashinfer to decide whether to use tensor core? |
I build from source and the version of flashinfer is 0.1.6
When I ran the demo of sglang, it gave me an error
I tried reconfiguring the environment but it didn't work.
Are there any other possible solutions?
The text was updated successfully, but these errors were encountered: