Skip to content

Latest commit

 

History

History
28 lines (18 loc) · 1.08 KB

README.md

File metadata and controls

28 lines (18 loc) · 1.08 KB

Transformers from scratch with PyTorch

A PyTorch implementation of the work "Attention is All You Need" with some goals:

Achieved:

  • An educational implementation (not a performant one)

Next steps:

  • Reproduce the results from the original paper

Requirements

  • Python 3.10.12
  • requirements.txt

Resources

These were the main resources I used to understand and implement the model.

Transformer Architecture:

Positional Encoding mechanism: