ConstraintLLM: A Neuro-Symbolic Framework for Industrial-Level Constraint Programming

📅 2025-10-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Automated modeling of constraint optimization problems (COPs) remains underdeveloped, and constraint programming (CP) modeling research lags behind operations research (OR). Method: We propose ConstraintLLM—the first large language model (LLM) specialized for CP tasks—integrating a constraint-aware retrieval module (CARM), a tree-of-thoughts (ToT) reasoning framework, and symbolic solvers to enable guided self-correction. It is built via multi-instruction supervised fine-tuning on open-source LLMs and evaluated on IndusCP, the first industrial-scale CP benchmark. Contribution/Results: ConstraintLLM achieves state-of-the-art (SOTA) solution accuracy on mainstream CP benchmarks and doubles modeling accuracy over baselines on IndusCP. It significantly improves reliability and generalization in formal COP modeling for industrial applications.

Technology Category

Application Category

📝 Abstract
Constraint programming (CP) is a crucial technology for solving real-world constraint optimization problems (COPs), with the advantages of rich modeling semantics and high solving efficiency. Using large language models (LLMs) to generate formal modeling automatically for COPs is becoming a promising approach, which aims to build trustworthy neuro-symbolic AI with the help of symbolic solvers. However, CP has received less attention compared to works based on operations research (OR) models. We introduce ConstraintLLM, the first LLM specifically designed for CP modeling, which is trained on an open-source LLM with multi-instruction supervised fine-tuning. We propose the Constraint-Aware Retrieval Module (CARM) to increase the in-context learning capabilities, which is integrated in a Tree-of-Thoughts (ToT) framework with guided self-correction mechanism. Moreover, we construct and release IndusCP, the first industrial-level benchmark for CP modeling, which contains 140 challenging tasks from various domains. Our experiments demonstrate that ConstraintLLM achieves state-of-the-art solving accuracy across multiple benchmarks and outperforms the baselines by 2x on the new IndusCP benchmark. Code and data are available at: https://github.com/william4s/ConstraintLLM.
Problem

Research questions and friction points this paper is trying to address.

Automates constraint programming modeling using large language models
Enhances in-context learning with constraint-aware retrieval mechanisms
Addresses industrial-level constraint optimization problems across diverse domains
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM fine-tuned for constraint programming modeling
Constraint-aware retrieval with guided self-correction
Industrial benchmark with 140 domain-specific tasks
🔎 Similar Papers
No similar papers found.
W
Weichun Shi
Hangzhou Institute for Advanced Study, UCAS, Hangzhou, China
M
Minghao Liu
University of Oxford, Oxford, UK
Wanting Zhang
Wanting Zhang
University of Science and Technology Beijing, Beijing, China
L
Langchen Shi
SKLCS and Key Laboratory of System Software, ISCAS, Beijing, China
Fuqi Jia
Fuqi Jia
Institute of Software Chinese Academy of Sciences
automated reasoningconstraint solvingcombinatorial optimization
Feifei Ma
Feifei Ma
Institute of Software, Chinese Academy of Sciences
Artificial Intelligence
J
Jian Zhang
SKLCS and Key Laboratory of System Software, ISCAS, Beijing, China