Profit Mirage: Revisiting Information Leakage in LLM-based Financial Agents

📅 2025-10-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper identifies a “profit illusion” in large language model (LLM)-based financial agents: high backtest returns stem from implicit information leakage within the model’s knowledge window, causing severe out-of-sample performance degradation. To address this, we propose FactFin—a causal-aware decision-making framework integrating four components: strategy code generation, retrieval-augmented generation, Monte Carlo tree search, and counterfactual simulation. By applying counterfactual perturbations, FactFin suppresses outcome memorization and steers the model toward learning causally grounded mechanisms. We further introduce FinLake-Bench—the first leakage-resistant benchmark for financial LLMs—that rigorously evaluates generalization under zero prior financial knowledge. Experiments demonstrate that FactFin significantly outperforms mainstream baselines in out-of-sample risk-adjusted returns (e.g., Sharpe ratio), validating the critical role of causal modeling in enhancing the robustness of LLM-driven financial decision-making.

Technology Category

Application Category

📝 Abstract
LLM-based financial agents have attracted widespread excitement for their ability to trade like human experts. However, most systems exhibit a "profit mirage": dazzling back-tested returns evaporate once the model's knowledge window ends, because of the inherent information leakage in LLMs. In this paper, we systematically quantify this leakage issue across four dimensions and release FinLake-Bench, a leakage-robust evaluation benchmark. Furthermore, to mitigate this issue, we introduce FactFin, a framework that applies counterfactual perturbations to compel LLM-based agents to learn causal drivers instead of memorized outcomes. FactFin integrates four core components: Strategy Code Generator, Retrieval-Augmented Generation, Monte Carlo Tree Search, and Counterfactual Simulator. Extensive experiments show that our method surpasses all baselines in out-of-sample generalization, delivering superior risk-adjusted performance.
Problem

Research questions and friction points this paper is trying to address.

Quantify information leakage in LLM financial agents
Develop leakage-robust benchmark for financial evaluation
Mitigate leakage via counterfactual causal learning framework
Innovation

Methods, ideas, or system contributions that make the work stand out.

Counterfactual perturbations prevent LLM memorization
Framework integrates retrieval-augmented generation with simulation
Monte Carlo Tree Search enhances causal reasoning capability
🔎 Similar Papers
No similar papers found.
X
Xiangyu Li
South China University of Technology, Guangzhou, China
Y
Yawen Zeng
ByteDance, Beijing, China
Xiaofen Xing
Xiaofen Xing
South China University of Technology
J
Jin Xu
South China University of Technology, Pazhou Lab, Guangzhou, China
Xiangmin Xu
Xiangmin Xu
South China University of Technology