AgentEvolver: How a 7B LLM Outperforms 14B Models with Self-Training

11 hours ago 高效码农

★AgentEvolver: A Self-Evolving Agent Framework That Writes Its Own Homework, Study Notes, and Report Card★ “ Can a large language model train itself to use tools in a brand-new environment without human-made datasets, dense reward functions, or brute-force sampling? Yes—AgentEvolver gives the model three “super-powers”: write the questions, remember the mistakes, and grade every step. The 7 B version outscores a 14 B baseline on two public benchmarks while using 60 % fewer tokens. 1. Why Most RL Pipelines for Agents Are Too Expensive Pain Point Symptom Cost No training tasks Engineers hand-write hundreds of multi-step questions $1–2 per label, …