Leveraging Encoder-only Large Language Models for Mobile App Review Feature Extraction

📅 2024-08-02
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
Mobile app reviews often exhibit low quality, high subjectivity, and substantial noise, posing significant challenges for feature extraction. To address this, we propose a supervised token classification approach for feature extraction. Our key contributions are threefold: (1) the first application of an encoder-only large language model (LLM) to this task; (2) a novel instance selection strategy based on uncertainty estimation, enabling efficient fine-tuning with reduced computational overhead; and (3) domain-adaptive pretraining, yielding an extended pre-trained model and a large-scale, high-quality crowdsourced annotation dataset. Extensive experiments demonstrate that our method achieves substantial improvements in both precision and recall over strong baselines, while simultaneously reducing fine-tuning costs—effectively balancing performance and efficiency.

Technology Category

Application Category

📝 Abstract
Mobile app review analysis presents unique challenges due to the low quality, subjective bias, and noisy content of user-generated documents. Extracting features from these reviews is essential for tasks such as feature prioritization and sentiment analysis, but it remains a challenging task. Meanwhile, encoder-only models based on the Transformer architecture have shown promising results for classification and information extraction tasks for multiple software engineering processes. This study explores the hypothesis that encoder-only large language models can enhance feature extraction from mobile app reviews. By leveraging crowdsourced annotations from an industrial context, we redefine feature extraction as a supervised token classification task. Our approach includes extending the pre-training of these models with a large corpus of user reviews to improve contextual understanding and employing instance selection techniques to optimize model fine-tuning. Empirical evaluations demonstrate that this method improves the precision and recall of extracted features and enhances performance efficiency. Key contributions include a novel approach to feature extraction, annotated datasets, extended pre-trained models, and an instance selection mechanism for cost-effective fine-tuning. This research provides practical methods and empirical evidence in applying large language models to natural language processing tasks within mobile app reviews, offering improved performance in feature extraction.
Problem

Research questions and friction points this paper is trying to address.

Enhances feature extraction from mobile app reviews
Redefines feature extraction as supervised token classification
Improves precision and recall in feature extraction
Innovation

Methods, ideas, or system contributions that make the work stand out.

Encoder-only large language models
Supervised token classification task
Extended pre-training with user reviews
🔎 Similar Papers
No similar papers found.
Q
Quim Motger
Universitat Politècnica de Catalunya, Department of Service and Information System Engineering, Barcelona, Spain
Alessio Miaschi
Alessio Miaschi
Researcher in Computational Linguistics, ItaliaNLP Lab @ CNR-ILC, Pisa
Natural Language ProcessingComputational LinguisticsLanguage ModelsDeep Learning
F
F. Dell’Orletta
Institute for Computational Linguistics “A. Zampolli” (ILC-CNR), ItaliaNLP Lab, Pisa, Italy
X
Xavier Franch
Universitat Politècnica de Catalunya, Department of Service and Information System Engineering, Barcelona, Spain
Jordi Marco
Jordi Marco
Associate Professor, Universitat Politècnica de Catalunya
Service Oriented ComputingNon-Functional RequirementsSoftware EngineeringComputer Graphics