Long Context
updated
MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via
Dynamic Sparse Attention
Paper
• 2407.02490
• Published
• 26
Can Few-shot Work in Long-Context? Recycling the Context to Generate
Demonstrations
Paper
• 2406.13632
• Published
• 5
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper
• 2406.15319
• Published
• 64
Make Your LLM Fully Utilize the Context
Paper
• 2404.16811
• Published
• 55
Megalodon: Efficient LLM Pretraining and Inference with Unlimited
Context Length
Paper
• 2404.08801
• Published
• 66
RULER: What's the Real Context Size of Your Long-Context Language
Models?
Paper
• 2404.06654
• Published
• 39
Long-context LLMs Struggle with Long In-context Learning
Paper
• 2404.02060
• Published
• 37
Evaluating Language Model Context Windows: A "Working Memory" Test and
Inference-time Correction
Paper
• 2407.03651
• Published
• 17
Associative Recurrent Memory Transformer
Paper
• 2407.04841
• Published
• 35
Long Code Arena: a Set of Benchmarks for Long-Context Code Models
Paper
• 2406.11612
• Published
• 25
Human-like Episodic Memory for Infinite Context LLMs
Paper
• 2407.09450
• Published
• 62
LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
Paper
• 2408.07055
• Published
• 68
RetrievalAttention: Accelerating Long-Context LLM Inference via Vector
Retrieval
Paper
• 2409.10516
• Published
• 43
HelloBench: Evaluating Long Text Generation Capabilities of Large
Language Models
Paper
• 2409.16191
• Published
• 41
Discovering the Gems in Early Layers: Accelerating Long-Context LLMs
with 1000x Input Token Reduction
Paper
• 2409.17422
• Published
• 25
L-CiteEval: Do Long-Context Models Truly Leverage Context for
Responding?
Paper
• 2410.02115
• Published
• 10