Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Wrapper with tests for gemmBatchedEx! #1975

Merged
merged 4 commits into from
Jun 29, 2023

Conversation

lpawela
Copy link
Contributor

@lpawela lpawela commented Jun 20, 2023

This introduces a wrapper for gemmBatchedEx!, allowing for batched mixed precision matrix multiplication. Based on the approach used for existing gemmEx! and gemm_batched!.

Copy link
Member

@maleadt maleadt left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks! Couple of minor comments.

I'm not really fond of these interfaces, they're low-level and ugly, but we don't have anything better right now.
Do you plan to use these directly, or to extend e.g. NNlib.batched_mul?

lib/cublas/wrappers.jl Outdated Show resolved Hide resolved
lib/cublas/wrappers.jl Outdated Show resolved Hide resolved
lib/cublas/wrappers.jl Outdated Show resolved Hide resolved
test/libraries/cublas.jl Show resolved Hide resolved
lpawela and others added 2 commits June 27, 2023 11:57
Co-authored-by: Tim Besard <tim.besard@gmail.com>
@lpawela
Copy link
Contributor Author

lpawela commented Jun 27, 2023

I'm not planning to add it NNlib, I'm using these directly. I will also add here (hopefuly today) the strided version.

@lpawela lpawela requested a review from maleadt June 27, 2023 23:43
@maleadt maleadt merged commit abd569e into JuliaGPU:master Jun 29, 2023
@lpawela lpawela deleted the lp/gemm-batched-ex branch March 20, 2024 13:21
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants