large-language-models
4 articles tagged with large-language-models
Google releases Gemma 4 family under Apache 2.0 license with 2B to 31B models
Google has released Gemma 4, a family of four open models ranging from 2B to 31B parameters, now available under the Apache 2.0 license for the first time. The 31B dense model ranks 3rd on the Arena AI Text Leaderboard, while the 26B mixture-of-experts variant ranks 6th, both outperforming significantly larger competitors. All models support multimodal inputs and are available on Hugging Face, Kaggle, and Ollama.
Xiaomi launches MiMo-V2-Pro with 1T parameters, matches Claude Opus on coding at 80% lower cost
Xiaomi shipped three AI models simultaneously designed to form a complete agent platform. MiMo-V2-Pro, a 1-trillion-parameter Mixture-of-Experts model with 42 billion active parameters per request, scores 78% on SWE-bench Verified and 81 points on ClawEval—nearly matching Claude Opus 4.6 while costing $1 per million input tokens versus $5 for Opus.
Meta pays News Corp up to $50M annually for AI training data in multi-year deal
Meta has committed to paying News Corp up to $50 million annually in a multi-year agreement for AI training data and content licensing. The deal represents Meta's continued strategy of securing high-quality publishing content for its AI models. The arrangement raises questions about the sustainability of individual content licensing deals versus industry-wide data standards.
Apple Intelligence generates stereotyped summaries across hundreds of millions of devices
Apple Intelligence, which automatically summarizes notifications and messages on hundreds of millions of devices, systematically generates stereotyped and hallucinated content according to an independent AI Forensics investigation. The analysis of over 10,000 AI-generated summaries reveals bias baked into the feature that pushes problematic assumptions to users unprompted.