-
-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add support for BFloat16 #41075
Comments
Since LLVM has support now, we need to figure out how to make the |
CUDA.jl's CUBLAS-based GEMM already works with BFloat16s like that, https://github.com/JuliaGPU/CUDA.jl/blob/f9f21dd3f80be936c30c44b051ffaaf599eac87e/lib/cublas/wrappers.jl#L868-L874. |
@Keno @maleadt @DhairyaLGandhi : When can the work for adding the avx512-bfloat16 instruction generation for x86 start? Similar to the work for float16 |
That would first require a way to target LLVM's |
This came up on the ML call today. BFloat16 has basically taken over Float16 as a low precision floating point representation and is now used in accelerating a number of ML ops to use lower precision element types. LLVM has also introduced the BFloat16 primitive. @Keno mentioned adding this support in the language generically would be possible and useful. Opening this here for folks to chime in with what needs to happen for this to come to fruition. BFloat16s.jl is already quite old at this point too.
Also cc @maleadt for his thoughts.
The text was updated successfully, but these errors were encountered: