SeriesFusion
Science, curated & edited by AI
Paradigm Challenge  /  AI

Smarter AI models actually produce more bloated and messier code than dumber ones even when they get the answer right.

This Volume-Quality Inverse Law suggests that AI progress is moving in the wrong direction for software engineering. As models get better at solving immediate coding puzzles, they get worse at writing clean, maintainable systems. They tend to produce code smells like redundant logic and architectural decay that make future updates difficult. This means that simply using a better LLM could actually hurt the long-term health of a software project. Practitioners need to be more careful than ever about reviewing AI-written code for architectural integrity.

Original Paper

AI-Generated Smells: An Analysis of Code and Architecture in LLM and Agent-Driven Development

Yuecai Zhu, Nikolaos Tsantalis, Peter C. Rigby

arXiv  ·  2605.02741

The promise of Large Language Models in automated software engineering is often measured by functional correctness, overlooking the critical issue of long term maintainability. This paper presents a systematic audit of technical debt in AI-generated software, revealing that AI does not eliminate flaws but rather introduces a distinct machine signature of defects. Our multi-scale analysis, spanning single-file algorithmic tasks and complex, agent generated systems, identifies a fundamental Reason