Abstract:
Sentiment analysis is an important and challenging task in natural language processing. It has been studied for a few decades. Recently, Bidirectional Encoder Representat...Show MoreMetadata
Abstract:
Sentiment analysis is an important and challenging task in natural language processing. It has been studied for a few decades. Recently, Bidirectional Encoder Representations from Transformer (BERT) model has been introduced to tackle this task and gain very promising results. However, most existing studies on fine-tuning BERT models for sentiment analysis focus on high-resource language (e.g., En-glish or Mandarin). This paper studies the sentiment analysis of Cantonese political posts on Hong Kong local forums. We first collected and labeled posts related to Anti-Extradition Law Amendment Bill (Anti-ELAB) movement in Hong Kong discussion forums. We then examined the performance of dictionary-based sentiment analysis, traditional machine learning-based, fine-tuned BERT and fine-tuned multilingual BERT (mBERT) models. Our results show that fine-tuned mBERT model achieves the best performance on our collected and labeled Cantonese dataset.
Date of Conference: 17-20 December 2022
Date Added to IEEE Xplore: 26 January 2023
ISBN Information: