🤖 AI Summary
This work addresses the challenge in Tree of Thoughts (ToT) frameworks of balancing deep exploration with computational efficiency during complex reasoning. The authors propose a lightweight, plug-and-play supervised Dynamic Search Trigger (DST) that eliminates reliance on costly large language model self-evaluation or fixed heuristic policies. DST employs a context-aware dynamic pruning mechanism, approximating greedy execution at straightforward reasoning steps while adaptively expanding the search beam at complex or uncertain nodes. This approach achieves efficient and scalable ToT reasoning, matching or surpassing standard ToT accuracy across mathematical, general, and complex logical reasoning benchmarks, while reducing computational overhead by 26%–75%.
📝 Abstract
While Large Language Models (LLMs) have advanced complex reasoning, prominent methods like the Tree of Thoughts (ToT) framework face a critical trade-off between exploration depth and computational efficiency. Existing ToT implementations often rely on heavyweight LLM-based self-evaluation or rigid heuristics for branch pruning, making them prohibitively expensive and inflexible for broad application. To address this, we introduce DST, an adaptable, plug-and-play predictor that serves as a lightweight, supervised heuristic to guide the ToT search process. Our predictor enables dynamic, context-aware pruning, allowing the search to proceed with near-greedy efficiency on simpler reasoning steps while adaptively expanding the search beam only when encountering uncertainty or task complexity. We evaluate our approach on a diverse suite of benchmarks spanning mathematical reasoning, general reasoning, and complex logical reasoning. Experimental results demonstrate that our method achieves accuracy competitive with or superior to strong baselines, including standard ToT, while reducing computational overhead by 26-75%. Our work effectively resolves the accuracy-efficiency trade-off in tree-based reasoning, transforming ToT from a resource-intensive technique into a scalable and practical paradigm for complex problem-solving in LLMs.