Efficient Streaming Language Models with Attention Sinks [ #3443
logikstate
started this conversation in
Ideas
Replies: 1 comment 3 replies
-
Hmm, either you misunderstood it or I did. https://github.com/mit-han-lab/streaming-llm#faq says that it doesn't actually increase context. As I understand it, this is basically #3377 - a more graceful way to do |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
New paper with example code claims huge context with minimal changes
https://github.com/mit-han-lab/streaming-llm
Beta Was this translation helpful? Give feedback.
All reactions