-
Attention Is All You Need
Paper • 1706.03762 • Published • 50 -
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
Paper • 2005.11401 • Published • 11 -
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 31 -
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper • 2205.14135 • Published • 11
Collections
Discover the best community collections!
Collections including paper arxiv:1706.03762
-
SIMPL: A Simple and Efficient Multi-agent Motion Prediction Baseline for Autonomous Driving
Paper • 2402.02519 • Published -
Mixtral of Experts
Paper • 2401.04088 • Published • 158 -
Optimal Transport Aggregation for Visual Place Recognition
Paper • 2311.15937 • Published -
GOAT: GO to Any Thing
Paper • 2311.06430 • Published • 14
-
Neural Machine Translation of Rare Words with Subword Units
Paper • 1508.07909 • Published • 4 -
Attention Is All You Need
Paper • 1706.03762 • Published • 50 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 16 -
Generating Wikipedia by Summarizing Long Sequences
Paper • 1801.10198 • Published • 3
-
Attention Is All You Need
Paper • 1706.03762 • Published • 50 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 12 -
Learning to summarize from human feedback
Paper • 2009.01325 • Published • 4 -
Training language models to follow instructions with human feedback
Paper • 2203.02155 • Published • 16