LongRoPE2: Near-Lossless LLM Context Window Scaling
Ning Shang, Li Lyna Zhang, Siyuan Wang, Gaokai Zhang, Gilsinia Lopez, Fan Yang, and 2 more authors
International Conference on Machine Learning, 2025
Large Language Models (LLMs) with extended context windows are essential for complex tasks. We present LongRoPE2, a novel method that extends LLM context windows to 128K tokens while retaining 98.5% short-context accuracy. Our approach introduces improved position encoding strategies that enable near-lossless context extension.