CARMA: Enhanced Compositionality in LLMs via Advanced Regularisation and Mutual Information Alignment

📅 2025-02-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) exhibit poor compositional generalization, struggling to systematically recombine learned semantics for interpreting novel compositions. To address this, we propose a lightweight, architecture- and data-agnostic intervention method. Our approach introduces three key innovations: (1) mutual information maximization regularization to enhance semantic consistency between inputs and internal representations; (2) joint intra-layer and inter-layer representation stability constraints to mitigate feature fragmentation induced by fine-tuning; and (3) an inverse dictionary modeling framework coupled with perturbation-robustness evaluation. Without compromising downstream task performance, our method significantly improves token-level semantic consistency and compositional reasoning accuracy across multiple benchmarks. It delivers consistent gains across diverse tasks and demonstrates strong cross-architecture transferability—validated on models including LLaMA and Qwen.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) struggle with compositional generalisation, limiting their ability to systematically combine learned components to interpret novel inputs. While architectural modifications, fine-tuning, and data augmentation improve compositionality, they often have limited adaptability, face scalability constraints, or yield diminishing returns on real data. To address this, we propose CARMA, an intervention that enhances the stability and robustness of compositional reasoning in LLMs while preserving fine-tuned performance. CARMA employs mutual information regularisation and layer-wise stability constraints to mitigate feature fragmentation, ensuring structured representations persist across and within layers. We evaluate CARMA on inverse dictionary modelling and sentiment classification, measuring its impact on semantic consistency, performance stability, and robustness to lexical perturbations. Results show that CARMA reduces the variability introduced by fine-tuning, stabilises token representations, and improves compositional reasoning. While its effectiveness varies across architectures, CARMA's key strength lies in reinforcing learned structures rather than introducing new capabilities, making it a scalable auxiliary method. These findings suggest that integrating CARMA with fine-tuning can improve compositional generalisation while maintaining task-specific performance in LLMs.
Problem

Research questions and friction points this paper is trying to address.

Enhances compositional generalization in LLMs
Reduces variability from fine-tuning
Improves robustness and stability in reasoning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Mutual information regularisation mitigates fragmentation
Layer-wise stability constraints ensure structured representations
CARMA enhances compositional reasoning in LLMs
🔎 Similar Papers
No similar papers found.
N
Nura Aljaafari
Department of Computer Science, University of Manchester, United Kingdom; King Faisal University, Al Hofuf, KSA
Danilo S. Carvalho
Danilo S. Carvalho
University of Manchester
Artificial IntelligenceNatural Language Processing
A
André Freitas
Department of Computer Science, University of Manchester, United Kingdom; Idiap Research Institute, Switzerland; National Biomarker Centre, CRUK-MI, Univ. of Manchester, United Kingdom