🤖 AI Summary
Existing approaches that manually translate domain theories into neural network architectures suffer from limited generality, verifiability, and scalability, often failing to guarantee strict alignment between models and prior knowledge. This work proposes a theory compiler that, for the first time, enables automatic mapping from formalized, typed domain theories to provably correct neural architectures. The method integrates formal methods, type systems, neural architecture design, and statistical learning theory through a universal theory language, a compositional compilation algorithm, and formal verification criteria, with large language models assisting in end-to-end compilation. The resulting architectures are theoretically guaranteed to achieve generalization performance comparable to or better than handcrafted designs while substantially reducing reliance on large training datasets.
📝 Abstract
Theory-guided machine learning has demonstrated that including authentic domain knowledge directly into model design improves performance, sample efficiency and out-of-distribution generalisation. Yet the process by which a formal domain theory is translated into architectural constraints remains entirely manual, specific to each domain formalism, and devoid of any formal correctness guarantee. This translation is non-transferable between domains, not verified, and does not scale. We propose the Theory Compiler: a system that accepts a typed, machine-readable domain theory as input and automatically produces an architecture whose function space is provably constrained to be consistent with that theory by construction, not by regularisation. We identify three foundational open problems whose resolution defines our research agenda: (1) designing a universal theory formalisation language with decidable type-checking; (2) constructing a compositionally correct compilation algorithm from theory primitives to architectural modules; and (3) establishing soundness and completeness criteria for formal verification. We further conjecture that compiled architectures match or exceed manually-designed counterparts in generalisation performance while requiring substantially less training data, a claim we ground in classical statistical learning theory. We argue that recent advances in formal machine learning theory, large language models, and the growth of an interdisciplinary research community have made this paradigm achievable for the first time.