LLM News

Every LLM release, update, and milestone.

Filtered by:llm✕ clear
research

New test-time training method improves LLM reasoning through self-reflection

Researchers propose TTSR, a test-time training framework where a single LLM alternates between Student and Teacher roles to improve its own reasoning. The method generates targeted variant questions based on analyzed failure patterns, showing consistent improvements across mathematical reasoning benchmarks without relying on unreliable pseudo-labels.

research

OSCAR: New RAG compression method achieves 2-5x speedup with minimal accuracy loss

Researchers have introduced OSCAR, a query-dependent compression method for Retrieval-Augmented Generation that speeds up inference 2-5x while preserving accuracy. Unlike traditional approaches, OSCAR compresses retrieved information dynamically at inference time rather than offline, eliminating storage overhead and enabling higher compression rates.

research

LaDiR uses latent diffusion to improve LLM reasoning beyond autoregressive limits

Researchers propose LaDiR, a framework that replaces traditional autoregressive decoding with latent diffusion models to improve LLM reasoning. The approach encodes reasoning steps into compressed latent representations and uses bidirectional attention to refine solutions iteratively, enabling parallel exploration of diverse reasoning paths.

2 min readvia arxiv.org
research

VeriStruct automates formal verification of Rust data structures with 99.2% function success rate

Researchers have introduced VeriStruct, a framework that extends AI-assisted formal verification from individual functions to complete data structure modules in Verus. The system successfully verified 128 of 129 functions (99.2%) across eleven Rust data structure modules by using a planner module to generate abstractions, type invariants, and proof code, with automatic error correction for Verus syntax.

2 min readvia arxiv.org