Implementation of the paper : Not all attention is needed - Gated Attention Network for Sequence Data (GA-Net)
The GAN model is implemented in Python using PyTorch framework. The GAN model is compared to attention and LSTM models for text classification problem on TREC and IMDb datasets.
There are two networks in the model:
- Backbone Network
- Auxiliary Network
Soft Attention gives some attention (low or high) to all the input tokens whereas gated attention network chooses the most important tokens to attend.
Visualization of probability for gate to be open for input token and the actual gated attention weight.