Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reuse GPU E-vecs #1648

Merged
merged 2 commits into from
Aug 29, 2024
Merged

Reuse GPU E-vecs #1648

merged 2 commits into from
Aug 29, 2024

Conversation

jeremylt
Copy link
Member

@jeremylt jeremylt commented Aug 22, 2024

This PR reuses active input e-vec buffers for output e-vecs where able. These are one of the biggest allocations of memory that the user doesn't see.

Adding this for AtPoints is a bit trickier. It is probably best as a follow-up

Note: manually tested with Ratel as well

@jeremylt jeremylt self-assigned this Aug 22, 2024
@jeremylt jeremylt force-pushed the jeremy/cuda-reuse-out branch 3 times, most recently from 95c081e to bcaca9e Compare August 26, 2024 19:16
@jeremylt
Copy link
Member Author

I did not add this change on the CPU side, as a 1 or 8 element block for the Opt backends is pretty negligible memory that doesn't seem worth the extra complexity in the code compared to the memory savings on the GPU side

@jeremylt jeremylt merged commit 71ed691 into main Aug 29, 2024
28 checks passed
@jeremylt jeremylt deleted the jeremy/cuda-reuse-out branch August 29, 2024 18:34
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants