🤖 AI Summary
Addressing the challenge of scarce labeled training data in dynamic combinatorial optimization (DCO), this paper proposes DyCO-GNN—the first fully training-data-free, unsupervised learning framework. DyCO-GNN leverages structural similarity across temporal graph snapshots, modeling dynamic topologies via graph neural networks and enabling cross-timestep knowledge transfer through unsupervised contrastive learning. It further incorporates a dynamic solution-space guidance mechanism to enhance computational efficiency. Evaluated on dynamic Max-Cut, Maximum Independent Set (MIS), and Traveling Salesman Problem (TSP) benchmarks, DyCO-GNN achieves speedups of 3–60× over state-of-the-art baselines while preserving solution quality. This significantly improves real-time responsiveness and resource efficiency, establishing a novel paradigm for DCO in zero-shot, label-free scenarios—where no historical instances or ground-truth labels are available.
📝 Abstract
We introduce DyCO-GNN, a novel unsupervised learning framework for Dynamic Combinatorial Optimization that requires no training data beyond the problem instance itself. DyCO-GNN leverages structural similarities across time-evolving graph snapshots to accelerate optimization while maintaining solution quality. We evaluate DyCO-GNN on dynamic maximum cut, maximum independent set, and the traveling salesman problem across diverse datasets of varying sizes, demonstrating its superior performance under tight and moderate time budgets. DyCO-GNN consistently outperforms the baseline methods, achieving high-quality solutions up to 3-60x faster, highlighting its practical effectiveness in rapidly evolving resource-constrained settings.