bird-of-paradise's picture
First commit
d9aa228 verified
---
license: mit
---
# Implementing Transformer from Scratch: A Step-by-Step Guide
This repository provides a detailed guide and implementation of the Transformer architecture from the ["Attention Is All You Need"](https://arxiv.org/abs/1706.03762) paper. The implementation focuses on understanding each component through clear code, comprehensive testing, and visual aids.
## Table of Contents
1. [Summary and Key Insights](#summary-and-key-insights)
2. [Implementation Details](#implementation-details)
- [Embedding and Positional Encoding](#embedding-and-positional-encoding)
- [Transformer Attention](#transformer-attention)
- [Feed-Forward Network](#feed-forward-network)
- [Transformer Decoder](#transformer-decoder)
- [Encoder-Decoder Stack](#encoder-decoder-stack)
- [Full Transformer](#full-transformer)
3. [Testing](#testing)
4. [Visualizations](#visualizations)
## Quick Start
View the complete implementation and tutorial in the [Jupyter notebook](Transformer_Implementation_Tutorial.ipynb).
## Summary and Key Insights
### Paper Reference
- ["Attention Is All You Need"](https://arxiv.org/abs/1706.03762) (Vaswani et al., 2017)
- Key sections:
- 3.1: Encoder and Decoder Stacks
- 3.2: Attention Mechanism
- 3.3: Position-wise Feed-Forward Networks
- 3.4: Embeddings and Softmax
- 3.5: Positional Encoding
- 5.4: Regularization (dropout strategy)
### Implementation Strategy
Breaking down the architecture into manageable pieces and gradually adding complexity:
1. Start with foundational components:
- Embedding + Positional Encoding
- Single-head self-attention
2. Build up attention mechanism:
- Extend to multi-head attention
- Add cross-attention capability
- Implement attention masking
3. Construct larger components:
- Encoder (self-attention + FFN)
- Decoder (masked self-attention + cross-attention + FFN)
4. Combine into final architecture:
- Encoder-Decoder stack
- Full Transformer with input/output layers
### Development Tips
1. Visualization and Planning:
- Draw out tensor dimensions on paper
- Sketch attention patterns and masks
- Map each component back to paper equations
- This helps catch dimension mismatches early!
2. Dimension Cheat Sheet:
- Input tokens: [batch_size, seq_len]
- Embeddings: [batch_size, seq_len, d_model]
- Attention matrices: [batch_size, num_heads, seq_len, seq_len]
- FFN hidden layer: [batch_size, seq_len, d_ff]
- Output logits: [batch_size, seq_len, vocab_size]
3. Common Pitfalls:
- Forgetting to scale dot products by √d_k
- Incorrect mask dimensions or application
- Missing residual connections
- Wrong order of layer norm and dropout
- Tensor dimension mismatches in attention
- Not handling padding properly
4. Performance Considerations:
- Memory usage scales with sequence length squared
- Attention computation is O(n²) with sequence length
- Balance between d_model and num_heads
- Trade-off between model size and batch size
## Implementation Details
### Embedding and Positional Encoding
This implements the input embedding and positional encoding from Section 3.5 of the paper. Key points:
- Embedding dimension can differ from model dimension (using projection)
- Positional encoding uses sine and cosine functions
- Scale embeddings by √d_model
- Apply dropout to the sum of embeddings and positional encodings
Implementation tips:
- Use `nn.Embedding` for token embeddings
- Store scaling factor as float during initialization
- Remember to expand positional encoding for batch dimension
- Add assertion for input dtype (should be torch.long)
### Transformer Attention
Implements the core attention mechanism from Section 3.2.1. Formula: Attention(Q,K,V) = softmax(QK^T/√d_k)V
Key points:
- Supports both self-attention and cross-attention
- Handles different sequence lengths for encoder/decoder
- Scales dot products by 1/√d_k
- Applies attention masking before softmax
Implementation tips:
- Use separate Q,K,V projections
- Handle masking through addition (not masked_fill)
- Remember to reshape for multi-head attention
- Keep track of tensor dimensions at each step
### Feed-Forward Network (FFN)
Implements the position-wise feed-forward network from Section 3.3: FFN(x) = max(0, xW₁ + b₁)W₂ + b₂
Key points:
- Two linear transformations with ReLU in between
- Inner layer dimension (d_ff) is typically 2048
- Applied identically to each position
Implementation tips:
- Use nn.Linear for transformations
- Remember to include bias terms
- Position-wise means same transformation for each position
- Dimension flow: d_model → d_ff → d_model
### Transformer Decoder
Implements decoder layer from Section 3.1, with three sub-layers:
- Masked multi-head self-attention
- Multi-head cross-attention with encoder output
- Position-wise feed-forward network
Key points:
- Self-attention uses causal masking
- Cross-attention allows attending to all encoder outputs
- Each sub-layer followed by residual connection and layer normalization
Key implementation detail for causal masking:
- Create causal mask using upper triangular matrix:
```python
mask = torch.triu(torch.ones(seq_len, seq_len), diagonal=1)
mask = mask.masked_fill(mask == 1, float('-inf'))
```
This creates a pattern where position i can only attend to positions ≤ i
Using -inf ensures zero attention to future positions after softmax
Visualization of mask for seq_len=5:\
[[0, -inf, -inf, -inf, -inf],\
[0, 0, -inf, -inf, -inf],\
[0, 0, 0, -inf, -inf],\
[0, 0, 0, 0, -inf],\
[0, 0, 0, 0, 0]]
Implementation tips:
- Order of operations matters (masking before softmax)
- Each attention layer has its own projections
- Remember to pass encoder outputs for cross-attention
Careful with mask dimensions in self and cross attention
### Encoder-Decoder Stack
Implements the full stack of encoder and decoder layers from Section 3.1.
Key points:
- Multiple encoder and decoder layers (typically 6)
- Each encoder output feeds into all decoder layers
- Maintains residual connections throughout the stack
Implementation tips:
- Use nn.ModuleList for layer stacks
- Share encoder outputs across decoder layers
- Maintain consistent masking throughout
- Handle padding masks separately from causal masks
### Full Transformer
Combines all components into complete architecture:
- Input embeddings for source and target
- Positional encoding
- Encoder-decoder stack
- Final linear and softmax layer
Key points:
- Handles different vocabulary sizes for source/target
- Shifts decoder inputs for teacher forcing
- Projects outputs to target vocabulary size
- Applies log softmax for training stability
Implementation tips:
- Handle start tokens for decoder input
- Maintain separate embeddings for source/target
- Remember to scale embeddings
- Consider sharing embedding weights with output layer
### Testing
Our implementation includes comprehensive tests for each component:
- Shape preservation through layers
- Masking effectiveness
- Attention pattern verification
- Forward/backward pass validation
- Parameter and gradient checks
See the notebook for detailed test implementations and results.
### Visualizations
The implementation includes visualizations of:
- Attention patterns
- Positional encodings
- Masking effects
- Layer connectivity
These visualizations help understand the inner workings of the transformer and verify correct implementation.
For detailed code and interactive examples, please refer to the complete implementation notebook.