🤖 AI Summary
This study addresses the limitation of AI explanation methods that rely on handcrafted templates and fail to effectively enhance human active learning performance. We propose LENS, the first framework integrating symbolic program synthesis with large language models (LLMs) to automate and scale the generation of natural-language explanations from logical rules—eliminating template dependency. Our neuro-symbolic approach extracts logical rules via logic programming, synthesizes structured intermediate representations programmatically, and leverages LLMs for explanation generation, augmented by multi-LLM adjudication and human evaluation to ensure fidelity and readability. Experiments show LENS-generated explanations significantly outperform both direct LLM prompting and template-based baselines in explanation quality; however, they do not yield statistically significant improvements in human learning outcomes on simple tasks—suggesting overly complex explanations may impose cognitive load. The core contribution is establishing the first template-free paradigm for generating pedagogically effective AI explanations.
📝 Abstract
Ultra Strong Machine Learning (USML) refers to symbolic learning systems that not only improve their own performance but can also teach their acquired knowledge to quantifiably improve human performance. In this work, we present LENS (Logic Programming Explanation via Neural Summarisation), a neuro-symbolic method that combines symbolic program synthesis with large language models (LLMs) to automate the explanation of machine-learned logic programs in natural language. LENS addresses a key limitation of prior USML approaches by replacing hand-crafted explanation templates with scalable automated generation. Through systematic evaluation using multiple LLM judges and human validation, we demonstrate that LENS generates superior explanations compared to direct LLM prompting and hand-crafted templates. To investigate whether LENS can teach transferable active learning strategies, we carried out a human learning experiment across three related domains. Our results show no significant human performance improvements, suggesting that comprehensive LLM responses may overwhelm users for simpler problems rather than providing learning support. Our work provides a solid foundation for building effective USML systems to support human learning. The source code is available on: https://github.com/lun-ai/LENS.git.