Breaking

GitHub will train Copilot models on user interaction data starting April 2026

GitHub will use Copilot interaction data from Free, Pro, and Pro+ plan users to train AI models starting April 24, 2026, unless users actively opt out. The policy does not affect Copilot Business and Enterprise customers. Data shared will include prompts, outputs, code snippets, filenames, and repository structures.

March 26, 2026

Latest News

All news →
0
research

Google's TurboQuant compression cuts LLM memory needs by 6x, sparks memory chip stock selloff

Google unveiled TurboQuant, a compression technique that reduces memory required to run large language models by six times by optimizing key-value cache storage. Memory chipmakers Samsung, SK Hynix, and Micron fell 5-6% on concern the efficiency breakthrough could reduce future chip demand. Analysts expect the decline reflects profit-taking rather than a fundamental shift, as more powerful models will eventually require more advanced hardware.

0
benchmarkOpenAI

ARC-AGI-3 benchmark: frontier AI models score below 1%, humans solve all 135 tasks

The ARC Prize Foundation released ARC-AGI-3, an interactive benchmark requiring AI agents to explore environments, form hypotheses, and execute plans without instructions. All 135 environments were solved by untrained humans, yet frontier models—including Gemini 3.1 Pro Preview (0.37%), GPT 5.4 (0.26%), Opus 4.6 (0.25%), and Grok-4.20 (0.00%)—scored below 1%.

0
researchApple

Apple's RubiCap model generates better image captions with 3-7B parameters than 72B competitors

Apple researchers developed RubiCap, a framework for training dense image captioning models that achieve state-of-the-art results at 2B, 3B, and 7B parameter scales. The 7B model outperforms models up to 72 billion parameters on multiple benchmarks including CapArena and CaptionQA, while the 3B variant matches larger 32B models, suggesting efficient dense captioning doesn't require massive scale.

2 min readvia 9to5mac.com
0
research

Google's TurboQuant cuts AI inference memory by 6x using lossless compression

Google Research unveiled TurboQuant, a lossless memory compression algorithm that reduces AI inference working memory (KV cache) by at least 6x without impacting model performance. The technology uses vector quantization methods called PolarQuant and an optimization technique called QJL. Findings will be presented at ICLR 2026.

0
model release

Google launches Lyria 3 Pro music generator, claims training data is rights-cleared

Google has released Lyria 3 Pro, its latest AI music generation model capable of creating tracks up to three minutes long with improved understanding of musical structure. The model is available through Gemini, Google Vids, Vertex AI, and Google AI Studio. Google claims the training data comes from sources it has contractual and legal rights to use.

2 min readvia the-decoder.com
0
product updateAmazon Web Services

Amazon Bedrock adds three video analysis workflows for multimodal understanding at scale

Amazon Bedrock has introduced three distinct video analysis workflows that leverage multimodal foundation models to extract insights from video content at scale. The approaches—frame-based, shot-based, and multimodal embedding—are designed for different use cases and cost-performance trade-offs, with open-source reference implementations available on GitHub.

0
product updateAmazon Web Services

Amazon Bedrock adds reinforcement fine-tuning with OpenAI-compatible APIs

Amazon Bedrock now enables reinforcement fine-tuning (RFT) across multiple model families including Amazon Nova, open-weight models like OpenAI's GPT-OSS 20B, and Qwen 3 32B. The service automates the end-to-end customization workflow using GRPO optimization, allowing models to learn from feedback on multiple responses rather than static training datasets, with support for OpenAI-compatible APIs.

0
model release

AI2 releases MolmoWeb, open web agent matching proprietary systems with 8B parameters

The Allen Institute for AI has released MolmoWeb, a fully open web agent that operates websites using only screenshots without access to source code. The 8B-parameter model achieves 78.2% success on WebVoyager—nearly matching OpenAI's o3 at 79.3%—while being trained on one of the largest public web task datasets ever released.

3 min readvia the-decoder.com
0
changelog

Google's Lyria 3 Pro extends AI music generation to 3-minute songs with structural control

Google released Lyria 3 Pro, an updated music generation model capable of creating full 3-minute songs—six times longer than the 30-second limit of its predecessor launched last month. The new version adds granular control over song structure, allowing users to specify intros, verses, choruses, and bridges. It's available now for paid Gemini users, enterprise customers, and developers via API.

0
product updateOpenAI

OpenAI shutters Sora video tool after Disney deal collapse, signaling shift to enterprise focus

OpenAI announced the shutdown of its Sora video generation app on Tuesday via an X post, just two days after publishing usage guidelines and following Disney's withdrawal from a proposed $1 billion investment deal. The move represents OpenAI's second major product discontinuation in recent months, after deprecating GPT-4o in January with two weeks' notice.

0
product update

Google's Gemini app now creates 3-minute songs with Lyria 3 Pro

Google announced Lyria 3 Pro, expanding the Gemini app's music generation capability from 30-second tracks to full 3-minute songs. The model improves structural understanding of musical composition, allowing users to prompt for specific elements like intros, verses, choruses, and bridges. Available now for Gemini subscribers with tier-based daily limits (10-50 tracks/day) and in Vertex AI, Google AI Studio, and the Gemini API for developers.

Latest Models

All →