Elsevier

Neurocomputing

Volume 272, 10 January 2018, Pages 258-269
Neurocomputing

Multi-modality weakly labeled sentiment learning based on Explicit Emotion Signal for Chinese microblog

https://doi.org/10.1016/j.neucom.2017.06.078Get rights and content

Abstract

Understanding the sentiments of users from cross media contents which contain texts and images is an important task for many social network applications. However, due to the semantic gap between cross media features and sentiments, machine learning methods need a lot of human labeled samples. Furthermore, for each kind of media content, it is necessary to constantly add a lot of new human labeled samples because of new expressions of sentiments. Fortunately, there are some emotion signals, like emoticons, which denote users’ emotions in cross media contents. In order to use these weakly labels to build a unified multi-modality sentiment learning framework, we propose an Explicit Emotion Signal (EES) based multi-modality sentiment learning approach which uses huge number of weakly labeled samples in sentiment learning. There are three advantages in our approach. Firstly, only a few human labeled samples are needed to reach the same performance which can be obtained by the traditional machine learning based sentiment prediction approaches. Secondly, this approach is flexible and can easily combine text and vision based sentiment learning through deep neural networks. Thirdly, because a lot of weakly labeled samples can be used in EES, trained model is more robust in different domain transfer. In this paper, firstly, we investigate the correlation between sentiments and emoticons and choose emoticons as the Explicit Emotion Signals in our approach; secondly, we build a two stages multi-modality sentiment learning framework based on Explicit Emotion Signals. Our experiment results show that our approach not only achieves the best performance but also only needs 3% and 43% training samples to obtain the same performance of Visual Geometry Group (VGG) model and Long Short-Term Memory (LSTM) model in images and texts, respectively.

Introduction

With the development of web social network applications, many people express their opinions by using cross media contents which contains texts, images and videos in Internet. In this situation, understanding the sentiment of users’ opinions is becoming more and more important. Generally speaking, sentiment analysis aims to determine the attitude of a speaker or a writer with respect to some topics or the overall contextual polarity of a document, such as positive, negative and neutral attitude. Recently, how to fully understand the sentiment from cross media yields an interesting research topic, cross media sentiment learning and analysis. One of the important issue for cross media sentiment learning in social network applications is that huge number of labeled samples are needed. There are three reasons which yield this problem. Firstly, unlike the common classification problems, sentiment is a kind of high level semantic information which is hard to achieve high performance directly from low level features, such as pixels and colors in images, words and phrases in texts. Therefore, many labeled samples are needed to learn the semantic space from raw data. Secondly, the problem becomes worse when we face the cross media content in social network applications. Because of the complementation of cross media content, many posts contain different sentiments in different media contents. To automatically learn sentiments from cross media contents, annotations for each kind of media content are needed. Thirdly, the way of expressing sentiment is changing over time which means that we need to constantly add a lot of new labeled samples to learn the change of sentiment expression.

Our goal is understanding the sentiments of users’ opinions. However, it is very difficult to obtain a large number of human labeled sentiment data. Fortunately, users often use emoticons in Chinese microblog, and those emoticons show users’ emotions. Therefore, we can collect a lot of weakly labeled samples which contain emotions. Here, weakly labeled sample means that the data only contains some emotion related symbols, such as emoticons, but without any sentiment labels which are the learning target. We think that an effective way of using weakly labeled samples will greatly alleviate the multi-modality sentiment learning problem. Although Hu et al. [1] used weakly labeled samples in the text based sentiment analysis, this approach is not suitable for cross media sentiment learning. Firstly, it used the orthogonal nonnegative matrix tri-factorization model which cannot handle the huge number of weakly labeled data to learn the latent sentiment space. Secondly, image content cannot be processed in the word content matrix. Therefore, we propose an Explicit Emotion Signal (EES) based approach which learns the EES from weakly labeled cross media samples from multi-modality view. Compared with state-of-the-art approaches, huge number of weakly labeled cross media contents can be used in our multi-modality sentiment learning framework and more robust features can be learned, and sentiment predication performance can be improved by using a few labeled samples.

In particular, we make the following contributions:

  • 1.

    We propose an Explicit Emotion Signal based cross media sentiment learning approach to use huge number of weakly labeled samples in social network applications, and it learns robust features which are not sensitive in different domain transfer.

  • 2.

    We implement a two stages multi-modality sentiment learning framework for both images and texts. This framework is flexible and can be transformed into a single-modality sentiment learning framework. The single-modality of the framework outperforms two state-of-the-art approaches, Long Short-Term Memory (LSTM) [2] and Visual Sentiment Ontology (VSO)[3], in text and image sentiment predication, respectively.

The remainder of this paper is organized as follows. Section 2 briefly describes the background and related work in sentiment analysis. Section 3 presents the concept of Explicit Emotion Signal and shows the statistical relationship between EES and sentiment. The algorithm and framework of EES based approach is detailed in Section 4. Experimental results on Sina Weibo datasets are given in Section 5. Section 6 draws conclusions and gives directions for future work.

Section snippets

Related work

Until now, the research works about sentiment analysis can be divided into two aspects: text based and vision based.

Traditional sentiment research works focus on text based sentiment analysis because words are the most common way of expressing opinion. According to the granularity of analysis, text based sentiment research works can be divided into three levels: document level [4], sentence level [5] and entity level [6]. Recently, many works focus on how to automatically mining sentiment word

Explicit Emotion Signal

In social media, people have used some symbols to denote their emotion in cross media content, such as adjective, adverb and emoticons. We call those symbols as Explicit Emotion Signals because those symbols contain the same characteristics. Firstly, users’ sentiments are not directly given by those symbols, it is necessary to combine those symbols with syntactic structure and context to understand users’ sentiments. Secondly, compared with some metaphors which need more knowledge to

Unified multi-modality sentiment learning framework

To solve the learning problem, we define a unified framework of multi-modality sentiment learning (Fig. 4), and the training algorithm is shown in Algorithm 1. It is a two stages sentiment learning framework. Firstly, huge number of weakly labeled samples are used to learn the middle level Explicit Emotion Signal features. In this stage, Explicit Emotion Signals are used as output layer to guide the middle level features learning. Secondly, a few labeled samples are used to learn the

Experiment

In this section, we have designed three kinds of experiments to verify the reasonableness and effectiveness of our approach. Because many cross media contents do not contain emoticons, to make our model more general, we only feed texts and images in the input layer and use emoticons as the output layer. In the following experiments: Firstly, we compared our Explicit Emotion Signal based multi-modality sentiment learning with other state-of-the-art sentiment learning approaches to show the

Conclusion

Cross media sentiment predication is a challenge work. How to learn good middle level features from cross media contents to understand sentiments is one of the key problems in this field. The state-of-the-art works, such as Visual Sentiment Ontology (VSO), Emotion words, etc. need a lot of labeled samples to learn middle level features. However, they are not applicable in the real-time social network applications. To solve that problem, we propose an Explicit Emotion Signal based cross media

Acknowlegment

This work is supported by the National Natural Science Foundation of China (Nos. 61402386, 61305061, 61502105, 61572409, 81230087 and 61571188), Open Fund Project of Fujian Provincial Key Laboratory of Information Processing and Intelligent Control (Minjiang University) (No. MJUKF201743), Education and Scientific Research Projects of young and middle-aged teachers in Fujian Province under Grant no. JA15075. Fujian Province 2011 Collaborative Innovation Center of TCM Health Management and

Dazhen Lin is current an assistant professor at the Department of Cognitive Science, School of Information Science and Engineering, Xiamen University. Her research interest is related to nature language processing and opinion mining.

References (34)

  • HuX. et al.

    Unsupervised sentiment analysis with emotional signals

    Proceedings of the 22nd International Conference on World Wide Web, WWW ’13

    (2013)
  • S. Hochreiter et al.

    Long short-term memory

    Neural Comput.

    (1997)
  • D. Borth et al.

    Large-scale visual sentiment ontology and detectors using adjective noun pairs

    Proceedings of the 21st ACM International Conference on Multimedia, MM ’13

    (2013)
  • PangB. et al.

    Thumbs up? Sentiment classification using machine learning techniques

    Proceedings of the ACL-02 Conference on Empirical Methods in Natural Language Processing, EMNLP ’02

    (2002)
  • ZhouL. et al.

    Unsupervised discovery of discourse relations for eliminating intra-sentence polarity ambiguities

    Proceedings of the Conference on Empirical Methods in Natural Language Processing, EMNLP ’11

    (2011)
  • HuM. et al.

    Mining and summarizing customer reviews

    Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’04

    (2004)
  • DingX. et al.

    A holistic lexicon-based approach to opinion mining

    Proceedings of the 2008 International Conference on Web Search and Data Mining, WSDM ’08

    (2008)
  • LiangJ. et al.

    Conr: a novel method for sentiment word identification

    Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, CIKM ’14

    (2014)
  • ZhangY. et al.

    Do users rate or review? Boost phrase-level sentiment labeling with review-level sentiment classification

    Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval, SIGIR ’14

    (2014)
  • F. Bravo-Marquez et al.

    From unlabelled tweets to twitter-specific opinion words

    Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’15

    (2015)
  • YuZ. et al.

    Featuring, detecting, and visualizing human sentiment in chinese micro-blog

    ACM Trans. Knowl. Discov. Data

    (2016)
  • C. Zhou, C. Sun, Z. Liu, F.C.M. Lau, A c-lstm Neural Network for Text Classification,...
  • A. Hogenboom et al.

    Exploiting emoticons in sentiment analysis

    Proceedings of the 28th Annual ACM Symposium on Applied Computing, SAC ’13

    (2013)
  • A. Agarwal et al.

    Sentiment analysis of twitter data

    Proceedings of the Workshop on Languages in Social Media, LSM ’11

    (2011)
  • ZhaoJ. et al.

    Moodlens: an emoticon-based sentiment analysis system for Chinese tweets

    Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’12

    (2012)
  • J. Machajdik et al.

    Affective image classification using features inspired by psychology and art theory

    Proceedings of the 18th ACM International Conference on Multimedia, MM ’10

    (2010)
  • V. Yanulevskaya et al.

    In the eye of the beholder: employing statistical analysis and eye tracking for analyzing abstract paintings

    Proceedings of the 20th ACM International Conference on Multimedia, MM ’12

    (2012)
  • Cited by (22)

    • Exploring user historical semantic and sentiment preference for microblog sentiment classification

      2021, Neurocomputing
      Citation Excerpt :

      With the proliferation of online reviews, ratings, and recommendations, mining user opinion has turned into a kind of virtual currency for many organisations looking to market their products, identifying new opportunities and managing their reputations. How to utilize machine learning technology to analyze the opinions of microblog posts has become one of the hotspots in the field of natural language research [1], and attracted considerable attention [2–5] from researchers in the past decade. Traditional sentiment classification technology is primarily based on exploiting sentiment lexicons or leveraging feature extraction techniques.

    • A cognitive brain model for multimodal sentiment analysis based on attention neural networks

      2021, Neurocomputing
      Citation Excerpt :

      Clustering algorithms are utilized to find common features group. For text modal, [39,40] construct a simplified Chinese sentiment dictionary based on micro-blog website and a Chinese car database is mentioned in [41] which is used to do knowledge-enhanced sentiment analysis. In [42], transfer learning is utilized that CNN and Bi-LSTM are used as base models.

    • Does government information release really matter in regulating contagion-evolution of negative emotion during public emergencies? From the perspective of cognitive big data analytics

      2020, International Journal of Information Management
      Citation Excerpt :

      Therefore, our work can fill in the research gap by conceptualizing social emotions in the under-presented emergency management scenario. Also, based on the work (Kušen, Cascavilla, Figl, Conti, & Strembeck, 2017; Lin, Li, Cao, Lv, & Ke, 2018; Stojanovski, Strezoski, Madjarov, Dimitrovski, & Chorbev, 2018; Wan, Jiang, Zhong, & Bian, 2013; Xu, Lin, & Zhao, 2008), our study conducts emotional lexicon using SO-PMI algorithm, and builds a fine-grained sentiment classifier with the aid of long- and short-term memory network technologies to improve the emotions classification performance. Second, the study explores the relationships between the netizens' emotion and the government information release and empirically tests how social negative emotions propagate through government information releasing under emergencies.

    • Modeling public mood and emotion: Stock market trend prediction with anticipatory computing approach

      2019, Computers in Human Behavior
      Citation Excerpt :

      Traditional approaches, such as semantic analysis and sentiment mining, were based on terms to predict whether the text content belongs to a positive or a negative sentiment. Lin (2018) mentioned that the textual sentiment mining research could be divided into three levels: text level, sentence level and document level. Recently, many studies aimed to find out emotional expressions in text data.

    View all citing articles on Scopus

    Dazhen Lin is current an assistant professor at the Department of Cognitive Science, School of Information Science and Engineering, Xiamen University. Her research interest is related to nature language processing and opinion mining.

    Lingxiao Li received the M.S. Degree in Computer Science and Technology in 2016 from Xiamen University, Xiamen, China. He is currently working in Alibaba Group.

    Donglin Cao is current an assistant professor at the Department of Cognitive Science, School of Information Science and Engineering, Xiamen University. His research interest is related to cross-media information retrieval, computer vision and nature language processing.

    Yanping Lv is received the Ph.D. degree in data mining from the University of Sherbrooke, Sherbrooke, QC, Canada, and the second Ph.D. degree in machine learning from Xiamen University, Fujian, China, in 2009. Her current research activity is related to machine learning, data mining and applications in ECG analysis, image classification, object tracking, etc.

    Xiao Ke received the Ph.D. degree in Cognitive Science from Xiamen University, Xiamen, China, in 2011. He is an assistant professor at Fuzhou University, China. His current research interests include machine learning and computer vision.

    View full text