This repository contains a PyTorch implementation of the Transformer model, based on the paper "Attention is All You Need" by Vaswani et al. (2017).
- Implements core Transformer components:
- Encoder-decoder architecture
- Multi-head attention
- Positional encoding
- Feed-forward network