Harnessing Agentic Evolution
2026-05-13 • Artificial Intelligence
Artificial IntelligenceMachine Learning
AI summaryⓘ
The authors present AEvo, a new way to improve programs or workflows by learning from past attempts and feedback over time. Instead of just making new guesses or candidates, AEvo edits the process that generates these guesses, allowing it to better use collected information. This approach helps AEvo guide long-term improvements more effectively than earlier methods. Their tests show AEvo works better than several existing techniques on various challenging tasks.
agentic evolutionmeta-agentfeedback loopslong-horizon optimizationcandidate generationprocess-level stateiterative searchprocedure editingevolutionary algorithmsopen-ended optimization
Authors
Jiayi Zhang, Yongfeng Gu, Jianhao Ruan, Maojia Song, Yiran Peng, Zhiguang Han, Jinyu Xiang, Zhitao Wang, Caiyin Yang, Yixi Ouyang, Bang Liu, Chenglin Wu, Yuyu Luo
Abstract
Agentic evolution has emerged as a powerful paradigm for improving programs, workflows, and scientific solutions by iteratively generating candidates, evaluating them, and using feedback to guide future search. However, existing methods are typically instantiated either as fixed hand-designed procedures that are modular but rigid, or as general-purpose agents that flexibly integrate feedback but can drift in long-horizon evolution. Both forms accumulate rich evidence over time, including candidates, feedback, traces, and failures, yet lack a stable interface for organizing this evidence and revising the mechanism that drives future evolution. We address this limitation by formulating agentic evolution as an interactive environment, where the accumulated evolution context serves as a process-level state. We introduce AEvo, a harnessed meta-editing framework in which a meta-agent observes this state and acts not by directly proposing the next candidate, but by editing the procedure or agent context that controls future evolution. This unified interface enables AEvo to steer both procedure-based and agent-based evolution, making accumulated evidence actionable for long-horizon search. Empirical evaluations on agentic and reasoning benchmarks show that AEvo outperforms five evolution baselines, achieving a 26 relative improvement over the strongest baseline. Across three open-ended optimization tasks, AEvo further outperforms four evolution baselines and achieves state-of-the-art performance under the same iteration budget.