A MIND for Reasoning: Meta-learning for In-context Deduction

πŸ“… 2025-05-20
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the poor out-of-distribution generalization of large language models (LLMs) on formal deductive reasoning tasks. We propose MIND, the first meta-learning-based in-context deduction framework, which integrates few-shot meta-instruction tuning, structured reasoning trace construction, dynamic context compression, and premise selection modeling to enable models to identify the minimal subset of premises required to derive a given hypothesis from a knowledge base. MIND achieves systematic generalization to unseen knowledge bases and inference rulesβ€”a capability previously unattained. Evaluated on 1.5B–7B parameter models, it significantly outperforms state-of-the-art larger models including GPT-4o and o3-mini, particularly under low-resource settings. Our approach bridges a critical gap between scalable instruction tuning and rigorous formal reasoning, establishing a new baseline for efficient, generalizable in-context deduction.

Technology Category

Application Category

πŸ“ Abstract
Large language models (LLMs) are increasingly evaluated on formal tasks, where strong reasoning abilities define the state of the art. However, their ability to generalize to out-of-distribution problems remains limited. In this paper, we investigate how LLMs can achieve a systematic understanding of deductive rules. Our focus is on the task of identifying the appropriate subset of premises within a knowledge base needed to derive a given hypothesis. To tackle this challenge, we propose Meta-learning for In-context Deduction (MIND), a novel few-shot meta-learning fine-tuning approach. The goal of MIND is to enable models to generalize more effectively to unseen knowledge bases and to systematically apply inference rules. Our results show that MIND significantly improves generalization in small LMs ranging from 1.5B to 7B parameters. The benefits are especially pronounced in smaller models and low-data settings. Remarkably, small models fine-tuned with MIND outperform state-of-the-art LLMs, such as GPT-4o and o3-mini, on this task.
Problem

Research questions and friction points this paper is trying to address.

Enhancing LLMs' generalization to out-of-distribution deductive problems
Systematic understanding of deductive rules in knowledge bases
Improving few-shot meta-learning for unseen inference tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Meta-learning fine-tuning for deduction
Improves generalization in small LMs
Outperforms GPT-4o on specific tasks
πŸ”Ž Similar Papers
No similar papers found.