🤖 AI Summary
Existing automated program synthesis research lacks standardized benchmark datasets in the ∀∃ logical format that support uncomputable symbolic constraints.
Method: We introduce the first ∀∃-logic benchmark suite tailored for deductive synthesis, formally modeling uncomputable symbolic restrictions, systematically extending classical instances, and incorporating a dynamic expansion mechanism. Our approach precisely encodes input-output specifications as ∀∃ formulas, integrating established benchmarks with newly constructed synthesis problems to ensure semantic rigor and task diversity.
Contribution/Results: The resulting dataset features a clear hierarchical structure and strong extensibility, filling a critical gap in standardized evaluation infrastructure for deductive synthesis. It provides a unified, reliable foundation for rigorously assessing both the theoretical capabilities and practical effectiveness of deductive synthesis algorithms—particularly their ability to handle quantified, non-computable constraints—thereby advancing principled evaluation in program synthesis research.
📝 Abstract
Program synthesis is the task of constructing a program conforming to a given specification. We focus on deductive synthesis, and in particular on synthesis problems with specifications given as $forallexists$-formulas, expressing the existence of an output corresponding to any input. So far there has been no canonical benchmark set for deductive synthesis using the $forallexists$-format and supporting the so-called uncomputable symbol restriction. This work presents such a data set, composed by complementing existing benchmarks by new ones. Our data set is dynamically growing and should motivate future developments in the theory and practice of automating synthesis.