FutureMind: Equipping Small Language Models with Strategic Thinking-Pattern Priors via Adaptive Knowledge Distillation

📅 2026-02-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of small language models in complex knowledge-intensive tasks, particularly their constrained capacity for structured reasoning and efficient retrieval. To overcome these challenges, the authors propose FutureMind, a novel framework that, for the first time, transfers strategic reasoning patterns—encompassing problem analysis, logical inference, strategic planning, and retrieval guidance—from large language models to smaller counterparts via adaptive knowledge distillation. FutureMind integrates these capabilities into a modular, dynamic reasoning pipeline that combines multi-hop question decomposition with a dynamic retrieval mechanism. The study further identifies cognitive discrepancies between teacher and student models as a key bottleneck in reasoning transfer. Evaluated on multi-hop QA benchmarks such as 2WikiMultihopQA and MuSiQue, FutureMind significantly outperforms strong baselines including Search-o1 and achieves state-of-the-art performance across small models of varying scales.

Technology Category

Application Category

📝 Abstract
Small Language Models (SLMs) are attractive for cost-sensitive and resource-limited settings due to their efficient, low-latency inference. However, they often struggle with complex, knowledge-intensive tasks that require structured reasoning and effective retrieval. To address these limitations, we propose FutureMind, a modular reasoning framework that equips SLMs with strategic thinking-pattern priors via adaptive knowledge distillation from large language models (LLMs). FutureMind introduces a dynamic reasoning pipeline composed of four key modules: Problem Analysis, Logical Reasoning, Strategy Planning, and Retrieval Guidance. This pipeline is augmented by three distinct retrieval paradigms that decompose complex queries into tractable subproblems, ensuring efficient and accurate retrieval execution. Extensive experiments on multi-hop QA benchmarks, including 2WikiMultihopQA, MuSiQue, Bamboogle, and Frames, demonstrate the superiority of FutureMind. It consistently outperforms strong baselines such as Search-o1, achieving state-of-the-art results under free training conditions across diverse SLM architectures and scales. Beyond empirical gains, our analysis reveals that the process of thinking-pattern distillation is restricted by the cognitive bias bottleneck between the teacher (LLMs) and student (SLMs) models. This provides new perspectives on the transferability of reasoning skills, paving the way for the development of SLMs that combine efficiency with genuine cognitive capability.
Problem

Research questions and friction points this paper is trying to address.

Small Language Models
structured reasoning
knowledge-intensive tasks
retrieval
cognitive capability
Innovation

Methods, ideas, or system contributions that make the work stand out.

adaptive knowledge distillation
strategic thinking-pattern priors
modular reasoning framework
cognitive bias bottleneck
small language models
🔎 Similar Papers
No similar papers found.
S
Shaoxiong Yang
MiLM Plus, Xiaomi Inc.
J
Junting Li
Department of Computer Science, Imperial College London
M
Mengyuan Zhang
MiLM Plus, Xiaomi Inc.
Chao Li
Chao Li
Baidu, Xiaomi
LLM-based AgentAI AgentAI Search
Wei Liu
Wei Liu
Alibaba Group
machine learningmedical image analysis
Jian Luan
Jian Luan
Toshiba, Microsoft, Xiaomi
LLMVLMTTSSinging Synthesis