LLM News

Every LLM release, update, and milestone.

Filtered by:swe-bench✕ clear
research

Half of AI code passing SWE-bench would be rejected by real developers, METR study finds

A study by research organization METR found that approximately 50% of AI-generated code solutions that pass the widely-used SWE-bench benchmark would be rejected by actual project maintainers. The finding exposes a significant gap between industry-standard code generation benchmarks and real-world code review standards.