Explaining Neural Networks in Preference Learning: a Post-hoc Inductive Logic Programming Approach

📅 2026-04-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the lack of interpretability in black-box neural networks for preference learning by proposing an approximation method that integrates Inductive Logic Programming via Answer Set Programming (ILASP) with Principal Component Analysis (PCA). PCA is employed to reduce the dimensionality of high-dimensional preference data, while ILASP introduces weak constraints to approximate the neural network’s behavior through global or local logical rules. Evaluated on a custom recipe preference dataset, the approach yields high-fidelity post-hoc explanations, significantly enhancing model transparency without compromising computational efficiency. The method thus offers a principled solution that balances accuracy and interpretability for neural preference models.
📝 Abstract
In this paper, we propose using Learning from Answer Sets to approximate black-box models, such as Neural Networks (NN), in the specific case of learning user preferences. We specifically explore the use of ILASP (Inductive Learning of Answer Set Programs) to approximate preference learning systems through weak constraints. We have created a dataset on user preferences over a set of recipes, which is used to train the NNs that we aim to approximate with ILASP. Our experiments investigate ILASP both as a global and a local approximator of the NNs. These experiments address the challenge of approximating NNs working on increasingly high-dimensional feature spaces while achieving appropriate fidelity on the target model and limiting the increase in computational time. To handle this challenge, we propose a preprocessing step that exploits Principal Component Analysis to reduce the dataset's dimensionality while keeping our explanations transparent. Under consideration for publication in Theory and Practice of Logic Programming (TPLP).
Problem

Research questions and friction points this paper is trying to address.

Preference Learning
Neural Networks
Explainability
Inductive Logic Programming
High-dimensional Data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Inductive Logic Programming
Preference Learning
Neural Network Interpretability
Answer Set Programming
Dimensionality Reduction
🔎 Similar Papers
No similar papers found.
D
Daniele Fossemò
Department of Information Engineering, Computer Science and Mathematics, University of L’Aquila, Italy
F
Filippo Mignosi
Department of Information Engineering, Computer Science and Mathematics, University of L’Aquila, Italy; ICAR-CNR, Palermo, Italy
G
Giuseppe Placidi
Department of Life, Health and Environmental Sciences, University of L’Aquila, Italy
Luca Raggioli
Luca Raggioli
University of Naples Federico II
Cognitive RoboticsSocial RoboticsHuman-Robot InteractionAI
Matteo Spezialetti
Matteo Spezialetti
DISIM - University of L'Aquila
Signal Analysis and ProcessingMachine LearningAffective Computing
Fabio Aurelio D'Asaro
Fabio Aurelio D'Asaro
University of Verona
Artificial IntelligenceLogicLogic ProgrammingProbability TheoryReasoning about Actions and