Long Context
updated
FLM-101B: An Open LLM and How to Train It with $100K Budget
Paper
• 2309.03852
• Published • 45
Extending LLMs' Context Window with 100 Samples
Paper
• 2401.07004
• Published • 16
LongAgent: Scaling Language Models to 128k Context through Multi-Agent
Collaboration
Paper
• 2402.11550
• Published • 19
The What, Why, and How of Context Length Extension Techniques in Large
Language Models -- A Detailed Survey
Paper
• 2401.07872
• Published • 2
A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts
Paper
• 2402.09727
• Published • 38
Data Engineering for Scaling Language Models to 128K Context
Paper
• 2402.10171
• Published • 25
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs
Miss
Paper
• 2402.10790
• Published • 42
LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning
Paper
• 2401.01325
• Published • 27
RetrievalAttention: Accelerating Long-Context LLM Inference via Vector
Retrieval
Paper
• 2409.10516
• Published • 43
Needle Threading: Can LLMs Follow Threads through Near-Million-Scale
Haystacks?
Paper
• 2411.05000
• Published • 22
Context Is What You Need: The Maximum Effective Context Window for Real
World Limits of LLMs
Paper
• 2509.21361
• Published • 11