Yedson54
's Collections
Long Context
updated
MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via
Dynamic Sparse Attention
Paper
•
2407.02490
•
Published
•
23
Can Few-shot Work in Long-Context? Recycling the Context to Generate
Demonstrations
Paper
•
2406.13632
•
Published
•
5
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper
•
2406.15319
•
Published
•
62
Make Your LLM Fully Utilize the Context
Paper
•
2404.16811
•
Published
•
52
Megalodon: Efficient LLM Pretraining and Inference with Unlimited
Context Length
Paper
•
2404.08801
•
Published
•
64
RULER: What's the Real Context Size of Your Long-Context Language
Models?
Paper
•
2404.06654
•
Published
•
34
Long-context LLMs Struggle with Long In-context Learning
Paper
•
2404.02060
•
Published
•
36
Evaluating Language Model Context Windows: A "Working Memory" Test and
Inference-time Correction
Paper
•
2407.03651
•
Published
•
16
Associative Recurrent Memory Transformer
Paper
•
2407.04841
•
Published
•
32
Long Code Arena: a Set of Benchmarks for Long-Context Code Models
Paper
•
2406.11612
•
Published
•
24
Human-like Episodic Memory for Infinite Context LLMs
Paper
•
2407.09450
•
Published
•
60
LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
Paper
•
2408.07055
•
Published
•
65
RetrievalAttention: Accelerating Long-Context LLM Inference via Vector
Retrieval
Paper
•
2409.10516
•
Published
•
40
HelloBench: Evaluating Long Text Generation Capabilities of Large
Language Models
Paper
•
2409.16191
•
Published
•
41
Discovering the Gems in Early Layers: Accelerating Long-Context LLMs
with 1000x Input Token Reduction
Paper
•
2409.17422
•
Published
•
24
L-CiteEval: Do Long-Context Models Truly Leverage Context for
Responding?
Paper
•
2410.02115
•
Published
•
10