Skip to content
This repository has been archived by the owner on Jan 3, 2023. It is now read-only.

Persistent RNNs #262

Closed
oleg-trott opened this issue Jun 20, 2016 · 4 comments
Closed

Persistent RNNs #262

oleg-trott opened this issue Jun 20, 2016 · 4 comments
Milestone

Comments

@oleg-trott
Copy link

Baidu found a way to get decent throughput in RNNs at very low batch sizes (like 4).

It would be interesting to see that in Neon/NervanaGPU

@jennifermyers
Copy link
Contributor

Hi Oleg,

Thanks for the note! We have added support for this internally and it is currently in test. We expect to release it in an upcoming version of neon - possibly not the imminent 1.5 release, but likely by 1.6.

@jennifermyers jennifermyers added this to the v1.6.0 milestone Jun 22, 2016
@oleg-trott
Copy link
Author

That was fast. I thought that paper just came out.

@thestew42
Copy link
Contributor

Greg gave a talk at GTC in April on persistent RNN kernels which we have been working to implement.

@jennifermyers
Copy link
Contributor

Good news! This is now expected to be part of the imminent v1.5.0 release.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

3 participants