|
--- |
|
license: mit |
|
--- |
|
|
|
# Implementing Transformer from Scratch: A Step-by-Step Guide |
|
|
|
This repository provides a detailed guide and implementation of the Transformer architecture from the ["Attention Is All You Need"](https://arxiv.org/abs/1706.03762) paper. The implementation focuses on understanding each component through clear code, comprehensive testing, and visual aids. |
|
|
|
## Table of Contents |
|
1. [Summary and Key Insights](#summary-and-key-insights) |
|
2. [Implementation Details](#implementation-details) |
|
- [Embedding and Positional Encoding](#embedding-and-positional-encoding) |
|
- [Transformer Attention](#transformer-attention) |
|
- [Feed-Forward Network](#feed-forward-network) |
|
- [Transformer Decoder](#transformer-decoder) |
|
- [Encoder-Decoder Stack](#encoder-decoder-stack) |
|
- [Full Transformer](#full-transformer) |
|
3. [Testing](#testing) |
|
4. [Visualizations](#visualizations) |
|
|
|
## Quick Start |
|
View the complete implementation and tutorial in the [Jupyter notebook](Transformer_Implementation_Tutorial.ipynb). |
|
|
|
## Summary and Key Insights |
|
|
|
### Paper Reference |
|
- ["Attention Is All You Need"](https://arxiv.org/abs/1706.03762) (Vaswani et al., 2017) |
|
- Key sections: |
|
- 3.1: Encoder and Decoder Stacks |
|
- 3.2: Attention Mechanism |
|
- 3.3: Position-wise Feed-Forward Networks |
|
- 3.4: Embeddings and Softmax |
|
- 3.5: Positional Encoding |
|
- 5.4: Regularization (dropout strategy) |
|
|
|
### Implementation Strategy |
|
Breaking down the architecture into manageable pieces and gradually adding complexity: |
|
|
|
1. Start with foundational components: |
|
- Embedding + Positional Encoding |
|
- Single-head self-attention |
|
|
|
2. Build up attention mechanism: |
|
- Extend to multi-head attention |
|
- Add cross-attention capability |
|
- Implement attention masking |
|
|
|
3. Construct larger components: |
|
- Encoder (self-attention + FFN) |
|
- Decoder (masked self-attention + cross-attention + FFN) |
|
|
|
4. Combine into final architecture: |
|
- Encoder-Decoder stack |
|
- Full Transformer with input/output layers |
|
|
|
### Development Tips |
|
1. Visualization and Planning: |
|
- Draw out tensor dimensions on paper |
|
- Sketch attention patterns and masks |
|
- Map each component back to paper equations |
|
- This helps catch dimension mismatches early! |
|
|
|
2. Dimension Cheat Sheet: |
|
- Input tokens: [batch_size, seq_len] |
|
- Embeddings: [batch_size, seq_len, d_model] |
|
- Attention matrices: [batch_size, num_heads, seq_len, seq_len] |
|
- FFN hidden layer: [batch_size, seq_len, d_ff] |
|
- Output logits: [batch_size, seq_len, vocab_size] |
|
|
|
3. Common Pitfalls: |
|
- Forgetting to scale dot products by √d_k |
|
- Incorrect mask dimensions or application |
|
- Missing residual connections |
|
- Wrong order of layer norm and dropout |
|
- Tensor dimension mismatches in attention |
|
- Not handling padding properly |
|
|
|
4. Performance Considerations: |
|
- Memory usage scales with sequence length squared |
|
- Attention computation is O(n²) with sequence length |
|
- Balance between d_model and num_heads |
|
- Trade-off between model size and batch size |
|
|
|
## Implementation Details |
|
|
|
### Embedding and Positional Encoding |
|
This implements the input embedding and positional encoding from Section 3.5 of the paper. Key points: |
|
- Embedding dimension can differ from model dimension (using projection) |
|
- Positional encoding uses sine and cosine functions |
|
- Scale embeddings by √d_model |
|
- Apply dropout to the sum of embeddings and positional encodings |
|
|
|
Implementation tips: |
|
- Use `nn.Embedding` for token embeddings |
|
- Store scaling factor as float during initialization |
|
- Remember to expand positional encoding for batch dimension |
|
- Add assertion for input dtype (should be torch.long) |
|
|
|
### Transformer Attention |
|
Implements the core attention mechanism from Section 3.2.1. Formula: Attention(Q,K,V) = softmax(QK^T/√d_k)V |
|
|
|
Key points: |
|
- Supports both self-attention and cross-attention |
|
- Handles different sequence lengths for encoder/decoder |
|
- Scales dot products by 1/√d_k |
|
- Applies attention masking before softmax |
|
|
|
Implementation tips: |
|
- Use separate Q,K,V projections |
|
- Handle masking through addition (not masked_fill) |
|
- Remember to reshape for multi-head attention |
|
- Keep track of tensor dimensions at each step |
|
|
|
### Feed-Forward Network (FFN) |
|
Implements the position-wise feed-forward network from Section 3.3: FFN(x) = max(0, xW₁ + b₁)W₂ + b₂ |
|
|
|
Key points: |
|
- Two linear transformations with ReLU in between |
|
- Inner layer dimension (d_ff) is typically 2048 |
|
- Applied identically to each position |
|
|
|
Implementation tips: |
|
- Use nn.Linear for transformations |
|
- Remember to include bias terms |
|
- Position-wise means same transformation for each position |
|
- Dimension flow: d_model → d_ff → d_model |
|
|
|
### Transformer Decoder |
|
Implements decoder layer from Section 3.1, with three sub-layers: |
|
- Masked multi-head self-attention |
|
- Multi-head cross-attention with encoder output |
|
- Position-wise feed-forward network |
|
|
|
Key points: |
|
- Self-attention uses causal masking |
|
- Cross-attention allows attending to all encoder outputs |
|
- Each sub-layer followed by residual connection and layer normalization |
|
|
|
Key implementation detail for causal masking: |
|
- Create causal mask using upper triangular matrix: |
|
```python |
|
mask = torch.triu(torch.ones(seq_len, seq_len), diagonal=1) |
|
mask = mask.masked_fill(mask == 1, float('-inf')) |
|
``` |
|
|
|
This creates a pattern where position i can only attend to positions ≤ i |
|
Using -inf ensures zero attention to future positions after softmax |
|
Visualization of mask for seq_len=5:\ |
|
[[0, -inf, -inf, -inf, -inf],\ |
|
[0, 0, -inf, -inf, -inf],\ |
|
[0, 0, 0, -inf, -inf],\ |
|
[0, 0, 0, 0, -inf],\ |
|
[0, 0, 0, 0, 0]] |
|
|
|
|
|
Implementation tips: |
|
- Order of operations matters (masking before softmax) |
|
- Each attention layer has its own projections |
|
- Remember to pass encoder outputs for cross-attention |
|
Careful with mask dimensions in self and cross attention |
|
|
|
### Encoder-Decoder Stack |
|
Implements the full stack of encoder and decoder layers from Section 3.1. |
|
Key points: |
|
- Multiple encoder and decoder layers (typically 6) |
|
- Each encoder output feeds into all decoder layers |
|
- Maintains residual connections throughout the stack |
|
|
|
Implementation tips: |
|
- Use nn.ModuleList for layer stacks |
|
- Share encoder outputs across decoder layers |
|
- Maintain consistent masking throughout |
|
- Handle padding masks separately from causal masks |
|
|
|
### Full Transformer |
|
Combines all components into complete architecture: |
|
- Input embeddings for source and target |
|
- Positional encoding |
|
- Encoder-decoder stack |
|
- Final linear and softmax layer |
|
|
|
Key points: |
|
- Handles different vocabulary sizes for source/target |
|
- Shifts decoder inputs for teacher forcing |
|
- Projects outputs to target vocabulary size |
|
- Applies log softmax for training stability |
|
|
|
Implementation tips: |
|
- Handle start tokens for decoder input |
|
- Maintain separate embeddings for source/target |
|
- Remember to scale embeddings |
|
- Consider sharing embedding weights with output layer |
|
|
|
### Testing |
|
Our implementation includes comprehensive tests for each component: |
|
|
|
- Shape preservation through layers |
|
- Masking effectiveness |
|
- Attention pattern verification |
|
- Forward/backward pass validation |
|
- Parameter and gradient checks |
|
|
|
See the notebook for detailed test implementations and results. |
|
|
|
### Visualizations |
|
The implementation includes visualizations of: |
|
|
|
- Attention patterns |
|
- Positional encodings |
|
- Masking effects |
|
- Layer connectivity |
|
|
|
These visualizations help understand the inner workings of the transformer and verify correct implementation. |
|
|
|
For detailed code and interactive examples, please refer to the complete implementation notebook. |
|
|