An overview of model uncertainty and variability in LLM-based sentiment analysis. Challenges, mitigation strategies and the role of explainability

📅 2025-04-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the Model Variability Problem (MVP)—a newly formalized challenge in LLM-based sentiment analysis arising from stochastic reasoning, prompt sensitivity, and training data bias. We systematically characterize how the temperature parameter governs output uncertainty and propose an interpretability-centered approach to enhance model trustworthiness. Methodologically, we integrate sensitivity analysis, prompt engineering evaluation, uncertainty quantification, and eXplainable AI (XAI) techniques to construct a unified assessment framework that jointly ensures stability, reproducibility, and interpretability. Our contributions include: (i) the first rigorous definition and empirical characterization of MVP; (ii) principled guidance on temperature tuning for uncertainty control; and (iii) a holistic evaluation protocol enabling robust, transparent sentiment analysis. Results demonstrate significant mitigation of classification inconsistency and output polarization, thereby facilitating deployment of trustworthy sentiment models in high-stakes domains such as finance, healthcare, and public policy.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have significantly advanced sentiment analysis, yet their inherent uncertainty and variability pose critical challenges to achieving reliable and consistent outcomes. This paper systematically explores the Model Variability Problem (MVP) in LLM-based sentiment analysis, characterized by inconsistent sentiment classification, polarization, and uncertainty arising from stochastic inference mechanisms, prompt sensitivity, and biases in training data. We analyze the core causes of MVP, presenting illustrative examples and a case study to highlight its impact. In addition, we investigate key challenges and mitigation strategies, paying particular attention to the role of temperature as a driver of output randomness and emphasizing the crucial role of explainability in improving transparency and user trust. By providing a structured perspective on stability, reproducibility, and trustworthiness, this study helps develop more reliable, explainable, and robust sentiment analysis models, facilitating their deployment in high-stakes domains such as finance, healthcare, and policymaking, among others.
Problem

Research questions and friction points this paper is trying to address.

Addresses inconsistency in LLM-based sentiment classification outcomes
Explores causes of model variability like prompt sensitivity and bias
Proposes strategies to enhance reliability and explainability in high-stakes domains
Innovation

Methods, ideas, or system contributions that make the work stand out.

Analyzes Model Variability Problem in LLMs
Explores temperature's role in output randomness
Emphasizes explainability for transparency and trust
🔎 Similar Papers
No similar papers found.
D
David Herrera-Poyatos
Department of Computer Science and Artificial Intelligence, Andalusian Institute of Data Science and Computational Intelligence (DaSCI), University of Granada, Spain.
C
Carlos Pel'aez-Gonz'alez
Department of Computer Science and Artificial Intelligence, Andalusian Institute of Data Science and Computational Intelligence (DaSCI), University of Granada, Spain.
Cristina Zuheros
Cristina Zuheros
University of Granada
Deep LearningSocial NetworksDecision MakingComputing with Words
A
Andr'es Herrera-Poyatos
Department of Computer Science and Artificial Intelligence, Andalusian Institute of Data Science and Computational Intelligence (DaSCI), University of Granada, Spain.
V
Virilo Tejedor
Department of Computer Science and Artificial Intelligence, Andalusian Institute of Data Science and Computational Intelligence (DaSCI), University of Granada, Spain.
Francisco Herrera
Francisco Herrera
Professor Computer Science and AI, DaSCI Research Institute, Granada University, Spain
Artificial IntelligenceComputational IntelligenceData ScienceTrustworthy AI
R
Rosana Montes
Department of Software Engineering, Andalusian Institute of Data Science and Computational Intelligence (DaSCI), University of Granada, Spain.