Transformer Implementation
Implementation of transformer encoder-decoder model based on paper Attention is all you need.
It includes (all built from scratch on pyTorch):
- Encoder and Decoder layers
- Multi-Head Attention and Scaled Dot-Product Attention
- Add & Norm layer
- Position-wise Feed-Forward Networks
- Sinusoidal Positional Encodings
All of them were implemented following paper..
Repository: https://github.com/rubzip/Transformers-Implementation