🤖 AI Summary
This work addresses the fundamental challenge of lacking formal verifiability in neural networks by introducing the novel paradigm of *Proof-Carrying Neural-Symbolic Code*, which unifies deep learning models and machine-checkable correctness proofs within a single executable code artifact—marking the first such integration. Methodologically, it synergistically combines formal verification (using Coq/Lean), neural program synthesis, differentiable symbolic execution, and automated theorem proving. We implement the first end-to-end prototype system capable of automatically generating and verifying machine-checkable proofs for small-scale neural-symbolic functions. Empirical evaluation demonstrates successful verification on multiple safety-critical micro-benchmarks. By bridging neural flexibility with symbolic rigor, this work establishes a principled foundation for building trustworthy AI systems that simultaneously possess strong learning capabilities and mathematically grounded safety guarantees.
📝 Abstract
This invited paper introduces the concept of"proof-carrying neuro-symbolic code"and explains its meaning and value, from both the"neural"and the"symbolic"perspectives. The talk outlines the first successes and challenges that this new area of research faces.