Contrastive Concept-Tree Search for LLM-Assisted Algorithm Discovery

๐Ÿ“… 2026-02-03
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the inefficiency of program search in large language model (LLM)-assisted algorithm discovery by proposing Contrastive Concept Tree Search (CCTS). CCTS introduces an explicit hierarchical concept representation, extracting concept trees from generated programs and leveraging contrastive learning to compute likelihood ratio scores between high- and low-performance solutions. These scores reweight parent selection during search, prioritizing effective concept combinations while avoiding misleading ones. Unlike approaches relying on LLMsโ€™ implicit algorithmic lineage, CCTS explicitly exploits โ€œconcepts to avoid,โ€ enhancing both search efficiency and interpretability. Evaluated on a benchmark of Erdล‘s-style combinatorial problems, CCTS significantly improves search performance, produces task-relevant interpretable concept trees, and demonstrates superior search dynamics in synthetic environments.

Technology Category

Application Category

๐Ÿ“ Abstract
Large language Model (LLM)-assisted algorithm discovery is an iterative, black-box optimization process over programs to approximatively solve a target task, where an LLM proposes candidate programs and an external evaluator provides task feedback. Despite intense recent research on the topic and promising results, how can the LLM internal representation of the space of possible programs be maximally exploited to improve performance is an open question. Here, we introduce Contrastive Concept-Tree Search (CCTS), which extracts a hierarchical concept representation from the generated programs and learns a contrastive concept model that guides parent selection. By reweighting parents using a likelihood-ratio score between high- and low-performing solutions, CCTS biases search toward useful concept combinations and away from misleading ones, providing guidance through an explicit concept hierarchy rather than the algorithm lineage constructed by the LLM. We show that CCTS improves search efficiency over fitness-based baselines and produces interpretable, task-specific concept trees across a benchmark of open Erd\H{o}s-type combinatorics problems. Our analysis indicates that the gains are driven largely by learning which concepts to avoid. We further validate these findings in a controlled synthetic algorithm-discovery environment, which reproduces qualitatively the search dynamics observed with the LLMs.
Problem

Research questions and friction points this paper is trying to address.

LLM-assisted algorithm discovery
program space representation
search efficiency
concept hierarchy
black-box optimization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Contrastive Concept-Tree Search
LLM-assisted algorithm discovery
concept hierarchy
program synthesis
contrastive learning
๐Ÿ”Ž Similar Papers
No similar papers found.
T
T. Leleu
1NTT Research, Sunnyvale, CA, USA; 2Stanford University, Palo Alto, USA
S
Sudeera Gunathilaka
3AIST, Tsukuba, Japan
F
Federico Ghimenti
2Stanford University, Palo Alto, USA
Surya Ganguli
Surya Ganguli
Associate Professor, Stanford University
NeurosciencePhysicsMachine Learning