๐ค AI Summary
Large language models (LLMs) struggle with complex, interdependent rule systems because they treat structured rules as unstructured text, thereby overlooking critical dependencies and introducing reasoning biases. To address this, we propose the Dynamic Adjudication Template (DAT), a three-stage reasoning framework that decouples inference into qualitative analysis, evidence collection, and adjudication. DAT integrates a placeholder-driven structured schema, chain-of-thought reasoning, and explicit rule validation to model rule dependencies explicitly and prevent error propagation. Empirical results demonstrate that DAT significantly improves accuracy for small-scale models on multi-level rule-based tasks. On multiple challenging rule-reasoning benchmarks, DAT outperforms standard chain-of-thought prompting and achieves performance comparable toโor even exceedingโthat of LLMs an order of magnitude larger in parameter count. This work establishes a novel paradigm for structured reasoning in lightweight models.
๐ Abstract
Large language models (LLMs) face significant challenges when processing complex rule systems, as they typically treat interdependent rules as unstructured textual data rather than as logically organized frameworks. This limitation results in reasoning divergence, where models often overlook critical rule dependencies essential for accurate interpretation. Although existing approaches such as Chain-of-Thought (CoT) reasoning have shown promise, they lack systematic methodologies for structured rule processing and are particularly susceptible to error propagation through sequential reasoning chains. To address these limitations, we propose the Dynamic Adjudication Template (DAT), a novel framework inspired by expert human reasoning processes. DAT structures the inference mechanism into three methodical stages: qualitative analysis, evidence gathering, and adjudication. During the qualitative analysis phase, the model comprehensively evaluates the contextual landscape. The subsequent evidence gathering phase involves the targeted extraction of pertinent information based on predefined template elements ([placeholder]), followed by systematic verification against applicable rules. Finally, in the adjudication phase, the model synthesizes these validated components to formulate a comprehensive judgment. Empirical results demonstrate that DAT consistently outperforms conventional CoT approaches in complex rule-based tasks. Notably, DAT enables smaller language models to match, and in some cases exceed, the performance of significantly larger LLMs, highlighting its efficiency and effectiveness in managing intricate rule systems.