Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Concurrent loop through each mini-batch during training #178

Merged
merged 3 commits into from
Jul 18, 2024

Conversation

rouson
Copy link
Contributor

@rouson rouson commented Jul 18, 2024

This PR lays the groundwork for parallelizing or offloading a significant part of the training algorithm to a GPU.

This commit allocates a pair_cost array for each mini-batch so
that each input/output pair's contribution to the cost function
sum can be stored in a separate element rather than keeping a
sequential running tally as the loop through the mini-batch
progresses.  This change lays the groundwork for making the
loop concurrent, which in turn lays the foundation for using
a Fortran 2023 concurrent reduction.  The calulaiton is currently
redundant with a running sum so that a match between the two can
be verified in an assertion.  The reundancy and the assertion will
be removed in a future commit.
@rouson rouson merged commit 5425151 into main Jul 18, 2024
6 checks passed
@rouson rouson deleted the concurrent-training branch July 18, 2024 05:47
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant