🤖 AI Summary
This study investigates large language models’ (LLMs) capacity to learn and generalize form-meaning mappings as defined by construction grammar, particularly exposing their systematic deficits in abstract reasoning—from concrete to schematic constructions—in natural language inference (NLI).
Method: We introduce ConTest-NLI, the first construction-driven, large-scale NLI benchmark (80K instances), covering eight English constructions. It employs template-based generation augmented with model-in-the-loop filtering to synthesize adversarial, schematic examples.
Contribution/Results: While state-of-the-art LLMs achieve 88% accuracy on natural data, performance drops sharply—down to 64%—on adversarial and schematic constructions. Construction-aware fine-tuning yields up to 9% absolute improvement, demonstrating both the necessity and feasibility of explicit constructional modeling. This work establishes the first scalable, construction-grammar–oriented evaluation framework for LLMs, revealing fundamental limitations in their understanding of abstract syntactic structure.
📝 Abstract
We probe large language models' ability to learn deep form-meaning mappings as defined by construction grammars. We introduce the ConTest-NLI benchmark of 80k sentences covering eight English constructions from highly lexicalized to highly schematic. Our pipeline generates diverse synthetic NLI triples via templating and the application of a model-in-the-loop filter. This provides aspects of human validation to ensure challenge and label reliability. Zero-shot tests on leading LLMs reveal a 24% drop in accuracy between naturalistic (88%) and adversarial data (64%), with schematic patterns proving hardest. Fine-tuning on a subset of ConTest-NLI yields up to 9% improvement, yet our results highlight persistent abstraction gaps in current LLMs and offer a scalable framework for evaluating construction-informed learning.