LLM News

Every LLM release, update, and milestone.

Filtered by:context-window✕ clear
research

New technique extends LLM context windows to 128K tokens without expensive retraining

Researchers propose a novel framework called SharedLLM that extends language model context windows from 8K to 128K tokens without costly continual pre-training. The method uses two stacked short-context models—one as a compressor, one as a decoder—with specialized tree-based information retrieval, achieving 2-3x inference speedups while maintaining competitive performance.

research

Neural Paging System Reduces LLM Context Management Complexity from O(N²) to O(N·K²)

A new research paper introduces Neural Paging, a hierarchical architecture that optimizes how LLMs manage their limited context windows by learning semantic caching policies. The approach reduces asymptotic complexity for long-horizon reasoning from O(N²) to O(N·K²) under bounded context window size K, addressing a fundamental bottleneck in deploying universal agents with external memory.