π€ AI Summary
Existing approaches to automatic agent workflow generation rely on homogeneous LLM pipelines and predefined operator libraries, resulting in high inference costs and substantial latency. This work proposes HyEvo, a novel framework that introduces a heterogeneous hybrid workflow architecture, decoupling deterministic computation from LLMs and synergistically combining LLM-driven semantic reasoning with rule-based execution in code nodes. HyEvo further incorporates an LLM-guided multi-island evolutionary algorithm, which iteratively refines both workflow structure and node logic through a reflection-generation mechanism enriched with execution feedback. Evaluated across multiple reasoning and programming benchmarks, HyEvo significantly outperforms state-of-the-art methods, achieving up to a 19Γ reduction in inference cost and a 16Γ decrease in execution latency.
π Abstract
Although agentic workflows have demonstrated strong potential for solving complex tasks, existing automated generation methods remain inefficient and underperform, as they rely on predefined operator libraries and homogeneous LLM-only workflows in which all task-level computation is performed through probabilistic inference. To address these limitations, we propose HyEvo, an automated workflow-generation framework that leverages heterogeneous atomic synthesis. HyEvo integrates probabilistic LLM nodes for semantic reasoning with deterministic code nodes for rule-based execution, offloading predictable operations from LLM inference and reducing inference cost and execution latency. To efficiently navigate the hybrid search space, HyEvo employs an LLM-driven multi-island evolutionary strategy with a reflect-then-generate mechanism, iteratively refining both workflow topology and node logic via execution feedback. Comprehensive experiments show that HyEvo consistently outperforms existing methods across diverse reasoning and coding benchmarks, while reducing inference cost and execution latency by up to 19$\times$ and 16$\times$, respectively, compared to the state-of-the-art open-source baseline.