Multi-Step Deductive Reasoning Over Natural Language: An Empirical Study on Out-of-Distribution Generalisation

📅 2022-07-28
🏛️ International Workshop on Neural-Symbolic Learning and Reasoning
📈 Citations: 12
Influential: 2
📄 PDF
🤖 AI Summary
This work addresses two key challenges in natural language multi-step deductive reasoning: poor out-of-distribution (OOD) generalization and skewed depth distribution in existing benchmarks. To tackle these, we propose IMA-GloVe-GA—a gated-attention-enhanced iterative memory network based on RNNs—integrating GloVe embeddings with explicit logical rule modeling, and introducing the first gated attention mechanism into the DeepLogic framework. Concurrently, we construct PARARULE-Plus, a large-scale, depth-balanced dataset that substantially alleviates the scarcity of deep-reasoning instances. Experiments demonstrate that our model surpasses both DeepLogic and RNN baselines on PARARULES and CONCEPTRULES. Under OOD settings involving rule reordering, it achieves higher accuracy than RoBERTa-Large. Moreover, on deep-reasoning tasks, PARARULE-Plus yields a 12.3% absolute accuracy improvement.
📝 Abstract
Combining deep learning with symbolic logic reasoning aims to capitalize on the success of both fields and is drawing increasing attention. Inspired by DeepLogic, an end-to-end model trained to perform inference on logic programs, we introduce IMA-GloVe-GA, an iterative neural inference network for multi-step reasoning expressed in natural language. In our model, reasoning is performed using an iterative memory neural network based on RNN with a gated attention mechanism. We evaluate IMA-GloVe-GA on three datasets: PARARULES, CONCEPTRULES V1 and CONCEPTRULES V2. Experimental results show DeepLogic with gated attention can achieve higher test accuracy than DeepLogic and other RNN baseline models. Our model achieves better out-of-distribution generalisation than RoBERTa-Large when the rules have been shuffled. Furthermore, to address the issue of unbalanced distribution of reasoning depths in the current multi-step reasoning datasets, we develop PARARULE-Plus, a large dataset with more examples that require deeper reasoning steps. Experimental results show that the addition of PARARULE-Plus can increase the model's performance on examples requiring deeper reasoning depths. The source code and data are available at https://github.com/Strong-AI-Lab/Multi-Step-Deductive-Reasoning-Over-Natural-Language.
Problem

Research questions and friction points this paper is trying to address.

Develops neural model for multi-step natural language reasoning
Improves out-of-distribution generalization in logical reasoning tasks
Addresses unbalanced reasoning depth distribution in existing datasets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines deep learning with symbolic logic reasoning
Uses iterative memory neural network with gated attention
Introduces PARARULE-Plus for deeper reasoning steps
🔎 Similar Papers
No similar papers found.
Qiming Bao
Qiming Bao
University of Auckland
Artificial IntelligenceNatural Language ProcessingLLMsReasoningNeurosymbolic AI
A
A. Peng
Strong AI Lab, NAOInstitute, Waipapa Taumata Rau - The University of Auckland, Auckland, New Zealand
T
Tim Hartill
Strong AI Lab, NAOInstitute, Waipapa Taumata Rau - The University of Auckland, Auckland, New Zealand
N
N. Tan
Strong AI Lab, NAOInstitute, Waipapa Taumata Rau - The University of Auckland, Auckland, New Zealand
Z
Zhenyun Deng
Department of Computer Science, University of Cambridge, Cambridge, The United Kingdom
M
M. Witbrock
Strong AI Lab, NAOInstitute, Waipapa Taumata Rau - The University of Auckland, Auckland, New Zealand
Jiamou Liu
Jiamou Liu
The University of Auckland
Social NetworksArtificial IntelligenceMachine Learning