🤖 AI Summary
Implicit aspect extraction in Aspect-Based Sentiment Analysis (ABSA) remains challenging due to the scarcity of real-world annotated data, particularly for emerging domains. Method: This study investigates the capability boundaries of large language models (LLMs) on implicit aspect extraction in a novel domain—sports—and introduces the first synthetic, generation-oriented dataset tailored for generative LLMs. It reformulates ABSA as a structured generation task and proposes a novel aspect–polarity pair evaluation metric. Systematic zero-shot and few-shot evaluations are conducted on open-source LLMs (e.g., LLaMA, Phi). Contribution/Results: Results demonstrate that LLMs possess preliminary capacity for implicit aspect identification, yet exhibit limited cross-domain generalization. The proposed metric significantly enhances comparability and interpretability of generated outputs. This work establishes a new paradigm for data construction and evaluation in low-resource ABSA, bridging the gap between generative modeling and fine-grained sentiment analysis.
📝 Abstract
This study examines the performance of Large Language Models (LLMs) in Aspect-Based Sentiment Analysis (ABSA), with a focus on implicit aspect extraction in a novel domain. Using a synthetic sports feedback dataset, we evaluate open-weight LLMs' ability to extract aspect-polarity pairs and propose a metric to facilitate the evaluation of aspect extraction with generative models. Our findings highlight both the potential and limitations of LLMs in the ABSA task.