Facilitating Long Context Understanding via Supervised Chain-of-Thought Reasoning

📅 2025-02-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the degradation of reasoning capability and poor interpretability in large language models (LLMs) when processing long-context inputs, this paper proposes a supervised chain-of-thought (CoT) enhancement framework. We first construct LongFinanceQA—the first synthetic financial-domain QA dataset featuring explicit, stepwise intermediate reasoning traces for long-context understanding. Second, we introduce Property-driven Agentic Inference (PAI), a novel inference framework that emulates human-like, multi-stage reasoning: property extraction → retrieval → summarization. Third, we systematically apply supervised CoT training to long-context tasks—marking the first such effort. Experiments demonstrate substantial improvements: GPT-4o-mini augmented with PAI achieves a +20.0% gain on the Loong benchmark; LLaMA-3.1-8B-Instruct improves by +24.6% on the Loong financial subset. Our approach significantly enhances both long-range reasoning accuracy and process interpretability.

Technology Category

Application Category

📝 Abstract
Recent advances in Large Language Models (LLMs) have enabled them to process increasingly longer sequences, ranging from 2K to 2M tokens and even beyond. However, simply extending the input sequence length does not necessarily lead to effective long-context understanding. In this study, we integrate Chain-of-Thought (CoT) reasoning into LLMs in a supervised manner to facilitate effective long-context understanding. To achieve this, we introduce LongFinanceQA, a synthetic dataset in the financial domain designed to improve long-context reasoning. Unlike existing long-context synthetic data, LongFinanceQA includes intermediate CoT reasoning before the final conclusion, which encourages LLMs to perform explicit reasoning, improving accuracy and interpretability in long-context understanding. To generate synthetic CoT reasoning, we propose Property-driven Agentic Inference (PAI), an agentic framework that simulates human-like reasoning steps, including property extraction, retrieval, and summarization. We evaluate PAI's reasoning capabilities by assessing GPT-4o-mini w/ PAI on the Loong benchmark, outperforming standard GPT-4o-mini by 20.0%. Furthermore, we fine-tune LLaMA-3.1-8B-Instruct on LongFinanceQA, achieving a 24.6% gain on Loong's financial subset.
Problem

Research questions and friction points this paper is trying to address.

Enhancing long-context understanding in LLMs
Supervised Chain-of-Thought reasoning integration
Synthetic dataset for financial domain reasoning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Supervised Chain-of-Thought reasoning
Property-driven Agentic Inference framework
LongFinanceQA synthetic dataset
🔎 Similar Papers
No similar papers found.
J
Jingyang Lin
University of Rochester, NY, United States; PAII Inc., CA, United States
A
Andy Wong
PAII Inc., CA, United States
T
Tian Xia
PAII Inc., CA, United States
S
Shenghua He
PAII Inc., CA, United States
H
Hui Wei
PAII Inc., CA, United States; University of California, Merced, CA, United States
Mei Han
Mei Han
Director of PAII Inc.
Computer VisionComputer GraphicsMachine Learning
J
Jiebo Luo
University of Rochester, NY, United States