LLM News

Every LLM release, update, and milestone.

Filtered by:open-source✕ clear
research

Researchers develop controllable full-duplex speech model trainable on 2,000 hours of data

Researchers have developed F-Actor, an instruction-following full-duplex conversational speech model that can be trained efficiently on 2,000 hours of data without large-scale pretraining. The model enables explicit control over speaker voice, conversation topic, backchanneling, interruptions, and dialogue initiation, addressing naturalness limitations in current spoken conversational systems.

research

MeanFlowSE enables single-step speech enhancement by learning mean velocity fields instead of instantaneous flows

Researchers introduced MeanFlowSE, a generative speech enhancement model that eliminates the computational bottleneck of multistep inference by learning average velocity over finite intervals rather than instantaneous velocity fields. The single-step approach achieves comparable quality to multistep baselines on VoiceBank-DEMAND while requiring substantially lower computational cost and no knowledge distillation.

research

xLLM: Open-source inference framework claims 2.2x vLLM throughput on Ascend accelerators

Researchers have released xLLM, an open-source Large Language Model inference framework designed for enterprise-scale serving. The framework claims to achieve up to 2.2x higher throughput than vLLM-Ascend when serving Qwen-series models under identical latency constraints, using a novel decoupled architecture that separates service scheduling from engine optimization.

2 min readvia arxiv.org
research

DeepXiv-SDK releases three-layer agentic interface for scientific literature access

DeepXiv-SDK introduces a three-layer agentic data interface designed to give LLM agents efficient, cost-aware access to scientific literature. The system transforms unstructured data into normalized JSON, offers retrieval tools via CLI, MCP, and Python SDK, and currently covers the complete arXiv corpus with daily synchronization.

2 min readvia arxiv.org
model release

Guide Labs open-sources Steerling-8B, an interpretable 8B parameter LLM

Guide Labs has open-sourced Steerling-8B, an 8 billion parameter language model built with a new architecture specifically designed to make the model's reasoning and actions easily interpretable. The release addresses a persistent challenge in AI development: understanding how large language models arrive at their outputs.

model release

Segmind releases SegMoE, a mixture-of-experts diffusion model for faster image generation

Segmind has released SegMoE, a mixture-of-experts (MoE) diffusion model designed to accelerate image generation while reducing computational overhead. The model applies MoE techniques traditionally used in large language models to the diffusion model architecture, enabling selective expert activation during inference.