FinZero: Launching Multi-modal Financial Time Series Forecast with Large Reasoning Model

📅 2025-09-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Financial time-series forecasting faces critical challenges including information loss due to standardization, rigid model architectures (fixed numbers of variables and window lengths), and insufficient interpretability and uncertainty quantification. To address these, we propose UARPO—a reinforcement learning-based fine-tuning framework enabling large language models to *adaptively assess prediction uncertainty*, dynamically select relevant variables, and adjust sliding window lengths. Furthermore, we introduce FVLDB, the first financial multimodal image–text dataset, and develop an end-to-end time-series reasoning pipeline grounded in vision–language foundation models. Experiments demonstrate that FinZero achieves a 13.48% accuracy gain over GPT-4o on high-confidence samples, significantly improving generalization, model interpretability, and deployment readiness in real-world financial applications.

Technology Category

Application Category

📝 Abstract
Financial time series forecasting is both highly significant and challenging. Previous approaches typically standardized time series data before feeding it into forecasting models, but this encoding process inherently leads to a loss of important information. Moreover, past time series models generally require fixed numbers of variables or lookback window lengths, which further limits the scalability of time series forecasting. Besides, the interpretability and the uncertainty in forecasting remain areas requiring further research, as these factors directly impact the reliability and practical value of predictions. To address these issues, we first construct a diverse financial image-text dataset (FVLDB) and develop the Uncertainty-adjusted Group Relative Policy Optimization (UARPO) method to enable the model not only output predictions but also analyze the uncertainty of those predictions. We then proposed FinZero, a multimodal pre-trained model finetuned by UARPO to perform reasoning, prediction, and analytical understanding on the FVLDB financial time series. Extensive experiments validate that FinZero exhibits strong adaptability and scalability. After fine-tuning with UARPO, FinZero achieves an approximate 13.48% improvement in prediction accuracy over GPT-4o in the high-confidence group, demonstrating the effectiveness of reinforcement learning fine-tuning in multimodal large model, including in financial time series forecasting tasks.
Problem

Research questions and friction points this paper is trying to address.

Addressing information loss in financial time series encoding
Overcoming limitations of fixed variable numbers and window lengths
Improving interpretability and uncertainty analysis in financial forecasting
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal pre-trained model for financial time series
Uncertainty-adjusted Group Relative Policy Optimization method
Reinforcement learning fine-tuning for prediction accuracy
🔎 Similar Papers
No similar papers found.
Y
Yanlong Wang
Tsinghua University, Pengcheng Laboratory
J
Jian Xu
Tsinghua University
F
Fei Ma
Guangming Laboratory
H
Hongkang Zhang
Tsinghua University
H
Hang Yu
Ant Group
T
Tiantian Gao
Tsinghua University
Y
Yu Wang
Pengcheng Laboratory, Southern University of Science and Technology
Haochen You
Haochen You
Columbia University
Generative AIMachine LearningStatistics
Shao-Lun Huang
Shao-Lun Huang
T-SIGS, Tsinghua University
Information TheoryMachine learning
D
Danny Dongning Sun
Pengcheng Laboratory
X
Xiao-Ping Zhang
Tsinghua University