🤖 AI Summary
Metaheuristics (MHs) for combinatorial optimization on social networks suffer from a lack of problem-specific prior knowledge, limiting their search efficacy. Method: This paper introduces the first large language model (LLM)-driven MH enhancement paradigm: an LLM serves as a structured pattern recognizer, extracting domain knowledge—e.g., community characteristics and influence propagation patterns—from problem descriptions and network topologies via customized prompt engineering; this knowledge dynamically guides neighborhood construction and operator selection within MHs. The approach requires no LLM fine-tuning, integrates seamlessly with mainstream MH frameworks, and yields interpretable, reproducible outputs. Contribution/Results: Evaluated on benchmark social network optimization tasks—including influence maximization and overlapping community detection—the method consistently outperforms existing machine learning–enhanced MHs, achieving an average 12.7% improvement in solution quality. The open-source toolkit OptiPattern is publicly released.
📝 Abstract
Since the rise of Large Language Models (LLMs) a couple of years ago, researchers in metaheuristics (MHs) have wondered how to use their power in a beneficial way within their algorithms. This paper introduces a novel approach that leverages LLMs as pattern recognition tools to improve MHs. The resulting hybrid method, tested in the context of a social network-based combinatorial optimization problem, outperforms existing state-of-the-art approaches that combine machine learning with MHs regarding the obtained solution quality. By carefully designing prompts, we demonstrate that the output obtained from LLMs can be used as problem knowledge, leading to improved results. Lastly, we acknowledge LLMs’ potential drawbacks and limitations and consider it essential to examine them to advance this type of research further. Our method can be reproduced using a tool available at: https://github.com/camilochs/optipattern.