🤖 AI Summary
This paper addresses the limited interpretability and weak structured argumentation capabilities of large language models (LLMs) in legal reasoning. To this end, it proposes the first three-level hierarchical tree-based modeling paradigm tailored for legal reasoning: rooted at the factum probandum (ultimate factual claim), with downward branching into evidentiary chains and implicit judicial experience nodes. Methodologically, the work introduces a crowdsourced, structured legal reasoning dataset; develops a legal-knowledge-enhanced LLM; designs tree-structured generation modeling; implements a multi-stage reasoning agent; and incorporates an evidence–experience alignment mechanism. Key contributions include: (1) releasing the first benchmark for legal reasoning trees; (2) enabling auditable, bias-mitigated structured causal reasoning; and (3) significantly improving argument accuracy and reviewability—thereby establishing a novel AI-assisted adjudication paradigm for smart courts.
📝 Abstract
While progress has been made in legal applications, law reasoning, crucial for fair adjudication, remains unexplored. We propose a transparent law reasoning schema enriched with hierarchical factum probandum, evidence, and implicit experience, enabling public scrutiny and preventing bias. Inspired by this schema, we introduce the challenging task, which takes a textual case description and outputs a hierarchical structure justifying the final decision. We also create the first crowd-sourced dataset for this task, enabling comprehensive evaluation. Simultaneously, we propose an agent framework that employs a comprehensive suite of legal analysis tools to address the challenge task. This benchmark paves the way for transparent and accountable AI-assisted law reasoning in the ``Intelligent Court''.