LexPam: Legal Procedure Awareness-Guided Mathematical Reasoning

📅 2025-04-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing legal large language models exhibit severe deficiencies in mathematical reasoning within legal contexts, particularly lacking task-specific modeling and evaluation for real-world applications such as compensation calculation. Method: We introduce LexNum, the first Chinese benchmark for legal mathematical reasoning, covering economic damages, work-related injury compensation, and traffic accident liability. We propose LexPam, a reinforcement learning algorithm guided by legal procedural awareness, integrating legal process modeling, multi-stage reasoning supervision, and domain-adaptive fine-tuning, optimized end-to-end via the PPO framework. Contribution/Results: Experiments demonstrate that LexPam significantly improves mathematical reasoning accuracy across all three compensation tasks, consistently outperforming state-of-the-art legal LLMs and general-purpose reasoning models. This work fills critical gaps in both legal-domain mathematical reasoning benchmarks and methodology.

Technology Category

Application Category

📝 Abstract
The legal mathematical reasoning ability of LLMs is crucial when applying them to real-world scenarios, as it directly affects the credibility of the LLM. While existing legal LLMs can perform general judicial question answering, their legal mathematical reasoning capabilities have not been trained. Open-domain reasoning models, though able to generate detailed calculation steps, do not follow the reasoning logic required for legal scenarios. Additionally, there is currently a lack of legal mathematical reasoning datasets to help validate and enhance LLMs' reasoning abilities in legal contexts. To address these issues, we propose the first Chinese legal Mathematical Reasoning Dataset, LexNum, which includes three common legal mathematical reasoning scenarios: economic compensation, work injury compensation, and traffic accident compensation. Based on LexNum, we tested the performance of existing legal LLMs and reasoning LLMs, and introduced LexPam, a reinforcement learning algorithm guided by legal procedural awareness to train LLMs, enhancing their mathematical reasoning abilities in legal scenarios. Experiments on tasks in the three legal scenarios show that the performance of existing legal LLMs and reasoning models in legal mathematical reasoning tasks is unsatisfactory. LexPam can enhance the LLM's ability in these tasks.
Problem

Research questions and friction points this paper is trying to address.

LLMs lack legal mathematical reasoning training for real-world credibility.
No dataset exists to validate legal math reasoning in LLMs.
Current models fail to follow legal scenario reasoning logic.
Innovation

Methods, ideas, or system contributions that make the work stand out.

First Chinese legal Mathematical Reasoning Dataset
Reinforcement learning with legal procedural awareness
Enhances LLMs' legal mathematical reasoning
🔎 Similar Papers
No similar papers found.
Kepu Zhang
Kepu Zhang
Renmin University of China
SearchLLMRecommendationLegal AI
Guofu Xie
Guofu Xie
Renmin University of China
Large Language ModelReinforcement Learning
W
Weijie Yu
University of International Business and Economics
M
Mingyue Xu
University of International Business and Economics
X
Xu Tang
University of International Business and Economics
Yaxin Li
Yaxin Li
University of International Business and Economics
J
Jun Xu
Gaoling School of Artificial Intelligence, Renmin University of China