Chain-of-Instructions: Compositional Instruction Tuning on Large Language Models

📅 2024-02-18
🏛️ arXiv.org
📈 Citations: 5
Influential: 0
📄 PDF
🤖 AI Summary
Existing large language models exhibit limited zero-shot generalization to multi-step compositional tasks—such as cross-lingual summarization—due to their reliance on single-step instruction paradigms. To address this, we propose the Chain-of-Instruction (CoI) paradigm, which explicitly models complex tasks as sequential chains of input-output subtasks, thereby enhancing end-to-end reasoning through stepwise decomposition. We formally define CoI for the first time, departing from conventional single-step instruction fine-tuning. Leveraging only existing instruction datasets, we construct CoI training samples and apply standard supervised fine-tuning (SFT), requiring no architectural modifications or reinforcement learning. Extensive experiments demonstrate that CoI-tuning consistently improves zero-shot generalization across compositional tasks—including long-chain reasoning, cross-lingual generation, and multi-hop question answering—with scalable gains. It significantly outperforms strong baselines across multiple benchmarks, establishing a new state-of-the-art in structured instruction learning.

Technology Category

Application Category

📝 Abstract
Fine-tuning large language models (LLMs) with a collection of large and diverse instructions has improved the model's generalization to different tasks, even for unseen tasks. However, most existing instruction datasets include only single instructions, and they struggle to follow complex instructions composed of multiple subtasks. In this work, we propose a novel concept of compositional instructions called chain-of-instructions (CoI), where the output of one instruction becomes an input for the next like a chain. Unlike the conventional practice of solving single instruction tasks, our proposed method encourages a model to solve each subtask step by step until the final answer is reached. CoI-tuning (i.e., fine-tuning with CoI instructions) improves the model's ability to handle instructions composed of multiple subtasks as well as unseen composite tasks such as multilingual summarization. Overall, our study find that simple CoI tuning of existing instruction data can provide consistent generalization to solve more complex, unseen, and longer chains of instructions.
Problem

Research questions and friction points this paper is trying to address.

Multi-step Tasks
Complex Task Handling
Multi-lingual Summarization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Command Chain (CoI)
Multi-step Task Execution
Sequential Instruction Processing
🔎 Similar Papers
No similar papers found.
S
S. Hayati
University of Minnesota
T
Taehee Jung
Amazon
T
Tristan Bodding-Long
Amazon
Sudipta Kar
Sudipta Kar
Principal Applied Scientist at Oracle Health AI
Artificial IntelligenceNatural Language ProcessingMachine LearningDeep Learning
A
A. Sethy
Grammarly
J
Joo-Kyung Kim
Amazon
Dongyeop Kang
Dongyeop Kang
University of Minnesota
Natural Language Processing