Achieving Semantic Consistency: Contextualized Word Representations for Political Text Analysis

📅 2024-12-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Static word embeddings (e.g., Word2Vec) exhibit semantic instability in political texts due to short-term lexical meaning shifts, undermining their reliability for longitudinal analysis. Method: We systematically compare BERT and Word2Vec across two dimensions—long-term semantic stability and short-term contextual sensitivity—using a 20-year corpus of *People’s Daily* political texts. We propose a time-segmented semantic similarity framework and introduce a novel quantitative metric for word vector stability. Contribution/Results: Our empirical analysis demonstrates that BERT’s contextualized embeddings significantly improve semantic stability over Word2Vec while preserving fine-grained sensitivity to policy-related lexical evolution. This finding challenges the conventional trade-off between stability and contextuality, offering a more robust and interpretable representation foundation for political text analysis requiring both temporal consistency and dynamic adaptability.

Technology Category

Application Category

📝 Abstract
Accurately interpreting words is vital in political science text analysis; some tasks require assuming semantic stability, while others aim to trace semantic shifts. Traditional static embeddings, like Word2Vec effectively capture long-term semantic changes but often lack stability in short-term contexts due to embedding fluctuations caused by unbalanced training data. BERT, which features transformer-based architecture and contextual embeddings, offers greater semantic consistency, making it suitable for analyses in which stability is crucial. This study compares Word2Vec and BERT using 20 years of People's Daily articles to evaluate their performance in semantic representations across different timeframes. The results indicate that BERT outperforms Word2Vec in maintaining semantic stability and still recognizes subtle semantic variations. These findings support BERT's use in text analysis tasks that require stability, where semantic changes are not assumed, offering a more reliable foundation than static alternatives.
Problem

Research questions and friction points this paper is trying to address.

Political Text Analysis
Word2Vec Instability
Semantic Nuance Capturing
Innovation

Methods, ideas, or system contributions that make the work stand out.

BERT
Word2Vec
Political Text Analysis
🔎 Similar Papers
No similar papers found.
Ruiyu Zhang
Ruiyu Zhang
The University of Hong Kong
Public ManagementOrganizationsBureaucracyComputational Social Science
L
Lin Nie
Department of Applied Social Sciences, The Hong Kong Polytechnic University
C
Ce Zhao
School of Science, Harbin Institute of Technology (Shenzhen)
Q
Qingyang Chen
Department of Applied Social Sciences, The Hong Kong Polytechnic University