Achieving Semantic Consistency: Contextualized Word Representations for Political Text Analysis

 0 Người đánh giá. Xếp hạng trung bình 0

Tác giả: Qingyang Chen, Lin Nie, Ruiyu Zhang, Ce Zhao

Ngôn ngữ: eng

Ký hiệu phân loại: 320.3 Comparative government

Thông tin xuất bản: 2024

Mô tả vật lý:

Bộ sưu tập: Metadata

ID: 205189

 Accurately interpreting words is vital in political science text analysis
  some tasks require assuming semantic stability, while others aim to trace semantic shifts. Traditional static embeddings, like Word2Vec effectively capture long-term semantic changes but often lack stability in short-term contexts due to embedding fluctuations caused by unbalanced training data. BERT, which features transformer-based architecture and contextual embeddings, offers greater semantic consistency, making it suitable for analyses in which stability is crucial. This study compares Word2Vec and BERT using 20 years of People's Daily articles to evaluate their performance in semantic representations across different timeframes. The results indicate that BERT outperforms Word2Vec in maintaining semantic stability and still recognizes subtle semantic variations. These findings support BERT's use in text analysis tasks that require stability, where semantic changes are not assumed, offering a more reliable foundation than static alternatives.Comment: 9 pages, 3 figures
Tạo bộ sưu tập với mã QR

THƯ VIỆN - TRƯỜNG ĐẠI HỌC CÔNG NGHỆ TP.HCM

ĐT: (028) 36225755 | Email: tt.thuvien@hutech.edu.vn

Copyright @2024 THƯ VIỆN HUTECH