LLM News

Every LLM release, update, and milestone.

0
model release

AI2 releases robotics models trained entirely in simulation, achieving zero-shot real-world transfer

AI2 has released MolmoSpaces and MolmoBot, robotics models trained exclusively in simulation that transfer directly to real robots without manual real-world data collection or fine-tuning. The approach eliminates months of teleoperated demonstrations typically required for simulation-trained robots. Both systems are open-source.

0
model release

Hume AI open-sources TADA: speech model 5x faster than rivals with zero hallucination

Hume AI has open-sourced TADA, a speech generation model that maps exactly one audio signal to each text token, achieving 5x faster processing than comparable systems. The model produced zero transcription hallucinations across 1,000+ test samples and runs on smartphones, available in 1B and 3B parameter versions under MIT license.

0
product updatePerplexity AI

Perplexity launches Personal Computer AI agent at $200/month for autonomous task handling

Perplexity AI has launched Personal Computer, a paid AI agent service priced at $200 per month that operates autonomously to handle emails, presentations, and application control. The service aims to provide continuous AI assistance for routine digital tasks without human intervention.

0
product updateNVIDIA

Nvidia to spend $26B on open-weight AI models, targeting Chinese competition and developer lock-in

An SEC filing reveals Nvidia plans to spend $26 billion on open-weight AI models over the next five years. The investment targets the open-source gap left by OpenAI, Meta, and Anthropic while countering the rise of Chinese open-source models and deepening developer dependence on Nvidia hardware.

0
model releaseNVIDIA

NVIDIA releases Nemotron-3-Super-120B, a 120B parameter model with latent MoE architecture

NVIDIA has released Nemotron-3-Super-120B-A12B-NVFP4, a 120-billion parameter text generation model featuring a latent Mixture-of-Experts (MoE) architecture. The model supports 8 languages including English, French, Spanish, Italian, German, Japanese, and Chinese, and is available on Hugging Face with 8-bit quantization support through NVIDIA's ModelOpt toolkit.

0
model releaseNVIDIA

NVIDIA releases Nemotron-3-Super-120B, a 120B parameter model with latent MoE architecture

NVIDIA has released Nemotron-3-Super-120B-A12B-BF16, a 120 billion parameter model designed for text generation and conversational tasks. The model employs a latent mixture-of-experts (MoE) architecture and supports multiple languages including English, French, Spanish, Italian, German, Japanese, and Chinese.

0
research

Half of AI code passing SWE-bench would be rejected by real developers, METR study finds

A study by research organization METR found that approximately 50% of AI-generated code solutions that pass the widely-used SWE-bench benchmark would be rejected by actual project maintainers. The finding exposes a significant gap between industry-standard code generation benchmarks and real-world code review standards.

0
research

AI agent compromised McKinsey's internal platform in 2 hours using SQL injection

An AI agent deployed by security firm Codewall gained full read and write access to McKinsey's internal AI platform Lilli within two hours without credentials or insider knowledge. The exploit used SQL injection, a decades-old vulnerability technique, to compromise a system serving over 43,000 employees for strategy work and client research.

0
product update

Qualcomm and Wayve partner to integrate physical AI into production vehicles

Qualcomm and Wayve announced a technical partnership to integrate Wayve's AI driving layer with Qualcomm's hardware platform for production-ready advanced driver assistance systems. The collaboration aims to accelerate autonomous vehicle innovation by combining hardware and software expertise.