🤖 AI Summary
This study addresses the neglect of annotators’ dynamic competence development in subjective annotation tasks by framing the annotation process itself as a skill-building mechanism, using social influence identification as a case study. Tracking the performance of 25 expert and non-expert annotators across 1,021 dialogue segments, the research integrates pre- and post-annotation comparisons, semi-structured interviews, self-assessment questionnaires, and large language model (LLM) training and evaluation based on the annotated data. Results reveal that active annotation significantly enhances annotators’ self-perceived competence and confidence, with experts exhibiting more pronounced gains. Crucially, the performance trajectory of LLMs trained on their annotations effectively mirrors this improvement, thereby validating the positive impact of the annotation process on downstream model efficacy.
📝 Abstract
Human data annotation, especially when involving experts, is often treated as an objective reference. However, many annotation tasks are inherently subjective, and annotators' judgments may evolve over time. This study investigates changes in the quality of annotators' work from a competence perspective during a process of social influence recognition. The study involved 25 annotators from five different groups, including both experts and non-experts, who annotated a dataset of 1,021 dialogues with 20 social influence techniques, along with intentions, reactions, and consequences. An initial subset of 150 texts was annotated twice - before and after the main annotation process - to enable comparison. To measure competence shifts, we combined qualitative and quantitative analyses of the annotated data, semi-structured interviews with annotators, self-assessment surveys, and Large Language Model training and evaluation on the comparison dataset. The results indicate a significant increase in annotators' self-perceived competence and confidence. Moreover, observed changes in data quality suggest that the annotation process may enhance annotator competence and that this effect is more pronounced in expert groups. The observed shifts in annotator competence have a visible impact on the performance of LLMs trained on their annotated data.