ITLC at SemEval-2026 Task 11: Normalization and Deterministic Parsing for Formal Reasoning in LLMs

📅 2026-03-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the susceptibility of large language models (LLMs) to content effects in multilingual logical reasoning, which often induces reasoning biases. To mitigate this issue, the authors propose a lightweight approach that transforms syllogisms into standardized logical forms through structured abstraction and leverages deterministic parsing to assess inferential validity—eliminating the need for complex fine-tuning or activation-based interventions. The method substantially reduces content-induced bias and enhances cross-lingual logical consistency. Evaluated on the SemEval-2026 Task 11 multilingual benchmark, it achieves top-five performance across all subtasks, demonstrating its effectiveness in strengthening the formal reasoning capabilities of LLMs without extensive model modifications.

Technology Category

Application Category

📝 Abstract
Large language models suffer from content effects in reasoning tasks, particularly in multi-lingual contexts. We introduce a novel method that reduces these biases through explicit structural abstraction that transforms syllogisms into canonical logical representations and applies deterministic parsing to determine validity. Evaluated on the SemEval-2026 Task 11 multilingual benchmark, our approach achieves top-5 rankings across all subtasks while substantially reducing content effects and offering a competitive alternative to complex fine-tuning or activation-level interventions.
Problem

Research questions and friction points this paper is trying to address.

content effects
reasoning tasks
large language models
multilingual contexts
logical reasoning
Innovation

Methods, ideas, or system contributions that make the work stand out.

structural abstraction
deterministic parsing
formal reasoning
content effects
canonical logical representation
🔎 Similar Papers
No similar papers found.
W
Wicaksono Leksono Muhamad
SEACrowd; Mantera Studio
J
Joanito Agili Lopo
SEACrowd; Kreasof AI
T
Tack Hwa Wong
SEACrowd; Universiti Teknologi PETRONAS
M
Muhammad Ravi Shulthan Habibi
SEACrowd; Universitas Indonesia
Samuel Cahyawijaya
Samuel Cahyawijaya
Cohere
Low-Resource NLPUnderrepresented LanguagesMultilingualCosslingualZero/Few-shot learning