Fin-R1: A Large Language Model for Financial Reasoning through Reinforcement Learning

📅 2025-03-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limitations of large language models (LLMs) in complex financial tasks—such as multi-step calculation, logical reasoning, and decision understanding—this paper introduces Fin-R1, a finance-specialized reasoning LLM. Methodologically, Fin-R1 employs a novel two-stage domain-specific training framework: Stage I constructs a high-quality, structured financial reasoning dataset via knowledge distillation from DeepSeek-R1; Stage II integrates supervised fine-tuning (SFT) and Proximal Policy Optimization (PPO)-based reinforcement learning to inject domain knowledge and enhance reasoning robustness. Empirical results demonstrate that Fin-R1 (7B) achieves state-of-the-art performance on FinQA and ConvFinQA, outperforming larger models. It also delivers substantial accuracy improvements in real-world financial tasks—including financial statement analysis and investment decision simulation—validating both the efficacy of domain-customized reasoning architectures and their generalization capability across diverse financial reasoning scenarios.

Technology Category

Application Category

📝 Abstract
Reasoning large language models are rapidly evolving across various domains. However, their capabilities in handling complex financial tasks still require in-depth exploration. In this paper, we introduce Fin-R1, a reasoning large language model specifically designed for the financial sector. Fin-R1 is built using a two-stage architecture, leveraging a financial reasoning dataset distilled and processed based on DeepSeek-R1. Through supervised fine-tuning (SFT) and reinforcement learning (RL) training, it demonstrates performance close to DeepSeek-R1 with a parameter size of 7 billion across a range of financial reasoning tasks. It achieves the state-of-the-art (SOTA) in the FinQA and ConvFinQA tasks between those LLMs in our evaluation, surpassing larger models in other tasks as well. Fin-R1 showcases strong reasoning and decision-making capabilities, providing solutions to various problems encountered in the financial domain. Our code is available at https://github.com/SUFE-AIFLM-Lab/Fin-R1.
Problem

Research questions and friction points this paper is trying to address.

Develops Fin-R1 for financial reasoning tasks
Enhances financial decision-making using reinforcement learning
Achieves SOTA in FinQA and ConvFinQA benchmarks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Two-stage architecture for financial reasoning
Supervised fine-tuning and reinforcement learning
State-of-the-art performance in financial tasks
🔎 Similar Papers
No similar papers found.
Z
Zhaowei Liu
Shanghai University of Finance and Economics
X
Xin Guo
Shanghai University of Finance and Economics
F
Fangqi Lou
Shanghai University of Finance and Economics
Lingfeng Zeng
Lingfeng Zeng
上海财经大学
大语言模型
J
Jinyi Niu
Fudan University
Z
Zixuan Wang
Shanghai University of Finance and Economics
J
Jiajie Xu
Shanghai University of Finance and Economics
W
Weige Cai
Shanghai University of Finance and Economics
Ziwei Yang
Ziwei Yang
Bioinformatics Center, Institute for Chemical Research, Kyoto University
BioinformaticsMachine LearningComputational BiologyBiomedical Data Science
X
Xueqian Zhao
Shanghai University of Finance and Economics
C
Chao Li
FinStep
S
Sheng Xu
FinStep
Dezhi Chen
Dezhi Chen
Beijing University of Posts and Telecommunications
UAV,Game theory,networks,Mean field
Y
Yun Chen
Shanghai University of Finance and Economics
Z
Zuo Bai
FinStep
L
Liwen Zhang
Shanghai University of Finance and Economics