Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
tatp22 authored Jul 6, 2020
1 parent adeaea3 commit 3973b39
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@

A practical implementation of the [Linformer paper](https://arxiv.org/pdf/2006.04768.pdf). This is attention with only linear complexity in n, allowing for very long sequence lengths (1mil+) to be attended to on modern hardware.

This repo is an [Attention Is All You Need](https://arxiv.org/pdf/1706.03762.pdf) style transformer, complete with an encoder and decoder (WIP) module. Check out how to use it below.
This repo is an [Attention Is All You Need](https://arxiv.org/pdf/1706.03762.pdf) style transformer, complete with an encoder and decoder (WIP) module. The novelty here is that now, one can make the attention heads linear. Check out how to use it below.

Visualization of the heads is also possible. To see more information, check out the Visualization section below.

Expand Down

0 comments on commit 3973b39

Please sign in to comment.