FORGE: Foundational Optimization Representations from Graph Embeddings

๐Ÿ“… 2025-08-27
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing learning-based combinatorial optimization approaches require separate model training for each problem distribution, suffering from poor generalization and high computational overhead. This paper proposes the first unsupervised pre-training framework for mixed-integer programming (MIP), built upon a vector-quantized graph autoencoder (VQ-GAE) to learn transferable instance representations: MIP structures are encoded as discrete, reusable symbols in a shared codebook, enabling unified modeling across diverse problem types. The framework operates without reliance on solver trajectories or labeled data, and supports multiple downstream tasksโ€”including variable warm-starting and cutting-plane generation. After fine-tuning on several MIP distributions, a single pre-trained Forge model significantly enhances the performance of commercial MIP solvers and demonstrates strong clustering capability on unseen instances.

Technology Category

Application Category

๐Ÿ“ Abstract
Combinatorial optimization problems are ubiquitous in science and engineering, yet learning-based approaches to accelerate their solution often require solving a large number of hard-to-solve optimization instances to collect training data, incurring significant computational overhead. Existing methods require training dedicated models for each problem distribution for each downstream task, severely limiting their scalability and generalization. In this work, we introduce Forge, a method of pre-training a vector-quantized graph autoencoder on a large and diverse collection of mixed-integer programming (MIP) instances in an unsupervised fashion without dependency on their solution. The vector quantization process creates discrete code assignments that act as a vocabulary to represent optimization instances. We evaluate our approach under both supervised and unsupervised settings. For the unsupervised setting, we demonstrate that Forge embeddings effectively differentiate and cluster unseen instances. For the supervised setting, we fine-tune Forge embeddings and show that a single model predicts both the variables for warm-starts and integrality gaps for cut-generation across multiple problem type distributions. Both predictions help improve performance of a state-of-the-art, commercial optimization solver. Finally, we release our code and pre-trained Forge weights to encourage further research and practical use of instance-level MIP embeddings at https://github.com/skadio/forge/
Problem

Research questions and friction points this paper is trying to address.

Pre-training graph embeddings for combinatorial optimization without solution dependency
Enhancing solver performance via warm-start variables and cut-generation predictions
Enabling generalization across diverse mixed-integer programming problem distributions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unsupervised pre-training on diverse MIP instances
Vector-quantized graph autoencoder for discrete representations
Single model predicts warm-starts and integrality gaps
๐Ÿ”Ž Similar Papers
No similar papers found.