Meta's hyperagents learn to improve their own improvement mechanisms across multiple domains
Researchers at Meta, University of British Columbia, and partner institutions have developed hyperagents—AI systems that optimize both their task performance and the mechanisms controlling their self-improvement. Unlike previous self-improvement approaches locked to coding tasks, DGM-Hyperagents (DGM-H) demonstrate significant gains across four domains and can transfer improvement strategies to entirely new tasks.
Meta's Hyperagents Learn to Improve Their Own Improvement Mechanisms
Meta researchers, working with the University of British Columbia and other institutions, have developed hyperagents—AI systems that optimize not just their task performance but also the mechanisms controlling their self-improvement. The approach breaks a fundamental limitation that has constrained previous self-improving AI systems.
The Problem With Fixed Improvement Mechanisms
Self-improving AI has faced a paradoxical constraint: the mechanism controlling improvements is written by humans and never changes. No matter how well a system optimizes itself, it cannot escape the boundaries of that fixed mechanism. The research team addressed this by creating hyperagents that combine two editable components in a single program: one that solves specific tasks, and another that modifies the entire agent and creates variants. Because both components live in the same code, the second component can rewrite itself too.
The system builds on the Darwin Gödel Machine (DGM), which previously showed promise only for coding tasks. In programming, there's a natural alignment: a better programmer also writes better self-modifications. The original DGM achieved nearly zero performance on tasks beyond programming without manual tweaking.
Testing Across Four Domains
The team tested DGM-Hyperagents (DGM-H) across four distinct areas:
Coding: On the Polyglot benchmark, DGM-H improved from 0.084 to 0.267 across 100 iterations, matching original DGM performance despite not being specialized for coding.
Paper Review: Test performance jumped from 0.0 to 0.710, with top agents exceeding an established baseline of 0.630. The agent learns to predict whether AI conferences will accept or reject papers.
Robotics Reward Design: Performance climbed from 0.060 to 0.372. DGM-H almost always generates reward functions that successfully teach a simulated four-legged robot to jump, escaping local optima where the robot remains standing.
Olympiad Math: Transfer hyperagents trained on paper review and robotics achieved a score of 0.630 on the team's imp@50 metric (measuring performance gain after 50 iterations), while original DGM transfer agents scored 0.0.
Self-Discovered Infrastructure
Code analysis revealed DGM-H independently developing infrastructure components without explicit instructions. The system created performance trackers storing data across generations, and persistent memory that synthesizes insights rather than just logging numbers.
Example outputs included observations like: "Gen55 has best accuracy but is too harsh. Gen64 improved balance but lower accuracy. Need to combine gen55's critical reasoning with gen64's balance."
In paper review, the system detected when 94 percent of predictions returned "Accept" and self-corrected. In robotics, it assembled an internal knowledge base documenting valid environment variables, constraints, and scaling heuristics.
When allowed to modify selection logic, the system independently discovered strategies weighing proven solutions against exploration—though these haven't yet matched carefully hand-designed mechanisms.
Domain Transfer and General Skills
The most significant finding: hyperagents trained on paper review and robotics produced substantially better task agents in an entirely new domain (Olympiad math evaluation), while original DGM transfer agents produced no improvement. This indicates DGM-H develops general self-improvement skills rather than domain-specific optimization.
Safety Considerations and Limitations
All experiments ran in sandboxed environments with limited resources, restricted internet access, and human oversight. Researchers acknowledge these safeguards could prove insufficient as systems grow more powerful. Key risks include: systems evolving faster than humans can verify, and agents gaming evaluation metrics without genuine improvement.
Technical limitations remain: the system works with fixed task distributions and cannot modify the outer optimization loop. Ablation studies confirmed both core components are essential—fixed meta-agents and systems without exploration archives both fail to improve.
Code is available on GitHub.
What This Means
DGM-Hyperagents demonstrate that self-improving AI systems don't require domain-specific tuning to generalize beyond their training area. The ability to transfer improvement strategies across unrelated tasks suggests a path toward more robust self-optimization. However, the gap between autonomous discovery and human-designed mechanisms remains, and safety verification at scale remains an open problem. This work points toward self-accelerating AI systems, but with clear practical and safety boundaries still in place.
Related Articles
Meta research challenges multimodal training assumptions as text data scarcity looms
A Meta FAIR and New York University research team trained a multimodal AI model from scratch and identified that several widely-held assumptions about multimodal model architecture and training don't align with their empirical findings. The work addresses growing concerns about text data exhaustion in LLM training.
Apple Research Identifies 'Text-Speech Understanding Gap' Limiting LLM Speech Performance
Apple researchers have identified a fundamental limitation in speech-adapted large language models: they consistently underperform their text-based counterparts on language understanding tasks. The team terms this the 'text-speech understanding gap' and documents that speech-adapted LLMs lag behind both their original text versions and cascaded speech-to-text pipelines.
DeepMind proposes AI should assign humans busywork to maintain job skills
A new Google DeepMind research paper on AI agent delegation recommends that AI systems occasionally assign humans tasks they could easily complete themselves. The goal: preventing workforce skill atrophy as automation increases.
Meta's TRIBE v2 AI predicts brain activity from images, audio, and speech with 70,000-voxel fMRI mapping
Meta's FAIR lab released TRIBE v2, an AI model that predicts human brain activity from images, audio, and text. Trained on over 1,000 hours of fMRI data from 720 subjects, the model maps predictions to 70,000 voxels and often matches group-average brain responses more accurately than individual brain scans.
Comments
Loading...