Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Delta neural networks inference #2129

Closed
casper2002casper opened this issue Nov 30, 2022 · 1 comment
Closed

Delta neural networks inference #2129

casper2002casper opened this issue Nov 30, 2022 · 1 comment

Comments

@casper2002casper
Copy link

casper2002casper commented Nov 30, 2022

Describe the potential feature

In scenarios where only part of the input layer is augmented, delta neural networks can be used to speed up inference by only updating the nodes affected by the change. The implementation of this principle has for example revolutionized chess engines (https://www.chessprogramming.org/NNUE), and has recently been implemented for CNNs https://github.com/facebookresearch/DeltaCNN.

Motivation

Could offer speedups in situations where the input layer is only partially updated, such as chess engines or combinatorial search methods.
Recent progress in CUDA sparse matrixes makes this interesting to evaluate.

Possible Implementation

Store hidden layer data from previous evaluation
Calculate input layer delta with previous evaluation and store in spare matrix
Update hidden layer data via sparse layer propagation

https://github.com/facebookresearch/DeltaCNN

@casper2002casper
Copy link
Author

Might be better as a seperate module

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant