RoME: Domain-Robust Mixture-of-Experts for MILP Solution Prediction across Domains

📅 2025-11-04
📈 Citations: 6
Influential: 0
📄 PDF
🤖 AI Summary
Existing learning-based methods for accelerating mixed-integer linear programming (MILP) solvers suffer from poor generalization and limited cross-domain transferability. To address this, we propose a task-embedding-driven initial solution prediction framework designed for cross-domain generalization. Our method introduces a Mixture-of-Experts architecture with dynamic instance routing and incorporates two-level distributionally robust optimization: inter-domain alignment and intra-domain perturbation-robust training—enhancing model adaptability to unseen problem distributions. To our knowledge, this is the first approach enabling efficient zero-shot transfer of a single model across diverse domains—including combinatorial optimization, scheduling, and network design. Evaluated on five benchmark domains, it achieves an average speedup of 67.7% over standard solvers. Moreover, on challenging real-world instances from MIPLIB, it significantly outperforms state-of-the-art methods, demonstrating both strong generalization capability and practical utility.

Technology Category

Application Category

📝 Abstract
Mixed-Integer Linear Programming (MILP) is a fundamental and powerful framework for modeling complex optimization problems across diverse domains. Recently, learning-based methods have shown great promise in accelerating MILP solvers by predicting high-quality solutions. However, most existing approaches are developed and evaluated in single-domain settings, limiting their ability to generalize to unseen problem distributions. This limitation poses a major obstacle to building scalable and general-purpose learning-based solvers. To address this challenge, we introduce RoME, a domain-Robust Mixture-of-Experts framework for predicting MILP solutions across domains. RoME dynamically routes problem instances to specialized experts based on learned task embeddings. The model is trained using a two-level distributionally robust optimization strategy: inter-domain to mitigate global shifts across domains, and intra-domain to enhance local robustness by introducing perturbations on task embeddings. We reveal that cross-domain training not only enhances the model's generalization capability to unseen domains but also improves performance within each individual domain by encouraging the model to capture more general intrinsic combinatorial patterns. Specifically, a single RoME model trained on three domains achieves an average improvement of 67.7% then evaluated on five diverse domains. We further test the pretrained model on MIPLIB in a zero-shot setting, demonstrating its ability to deliver measurable performance gains on challenging real-world instances where existing learning-based approaches often struggle to generalize.
Problem

Research questions and friction points this paper is trying to address.

Addressing poor generalization of learning-based MILP solvers across domains
Developing domain-robust framework using mixture-of-experts architecture
Enhancing cross-domain performance through distributionally robust optimization training
Innovation

Methods, ideas, or system contributions that make the work stand out.

Domain-Robust Mixture-of-Experts framework for MILP
Dynamic routing to experts using learned task embeddings
Two-level distributionally robust optimization strategy
🔎 Similar Papers
No similar papers found.
T
Tianle Pu
Laboratory for Big Data and Decision, College of Systems Engineering, National University of Defense Technology
Z
Zijie Geng
MoE Key Laboratory of Brain-inspired Intelligent Perception and Cognition, University of Science and Technology of China
H
Haoyang Liu
MoE Key Laboratory of Brain-inspired Intelligent Perception and Cognition, University of Science and Technology of China
Shixuan Liu
Shixuan Liu
National University of Defense Technology
Knowledge ReasoningDomain GeneralizationCausal InferenceData Engineering
J
Jie Wang
MoE Key Laboratory of Brain-inspired Intelligent Perception and Cognition, University of Science and Technology of China
Li Zeng
Li Zeng
Peking University
LLM training and inferenceVector ComputingGraph Computing
C
Chao Chen
Laboratory for Big Data and Decision, College of Systems Engineering, National University of Defense Technology
Changjun Fan
Changjun Fan
Associate Professor, National University of Defense Technology
graph neural networkcombinatorial optimizationreinforcement learning