Abstract
The current decade has witnessed the remarkable developments in the field of artificial intelligence, and the revolution of deep learning has transformed the whole artificial intelligence industry. Eventually, deep learning techniques have become essential components of any model in today’s computational world. Nevertheless, deep learning techniques promise a high degree of automation with generalized rule extraction for both text and sentiment classification tasks. This article aims to provide an empirical study on various deep neural networks (DNN) used for sentiment classification and its applications. In the preliminary step, the research carries out a study on several contemporary DNN models and their underlying theories. Furthermore, the performances of different DNN models discussed in the literature are estimated through the experiments conducted over sentiment datasets. Following this study, the effect of fine-tuning various hyperparameters on each model’s performance is also examined. Towards a better comprehension of the empirical results, few simple techniques from data visualization have been employed. This empirical study ensures deep learning practitioners with insights into ways to adapt stable DNN techniques for many sentiment analysis tasks.
Similar content being viewed by others
References
Akhtar MS, Gupta D, Ekbal A, Bhattacharyya P (2017) Feature selection and ensemble construction: a two-step method for aspect based sentiment analysis. Knowl Based Syst 125:116–135
Akhtar MS, Ekbal A, Narayan S, Singh V (2018) No, that never happened!! investigating rumors on twitter. IEEE Intell Syst 33(5):8–15
Akhtar MS, Ghosal D, Ekbal A, Bhattacharyya P, Kurohashi S (2019) All-in-one: emotion, sentiment and intensity prediction using a multi-task ensemble framework. IEEE Trans Affect Comput. https://doi.org/10.1109/TAFFC.2019.2926724
Akhtar MS, Ekbal A, Cambria E (2020) How intense are you? Predicting intensities of emotions and sentiments using stacked ensemble. IEEE Comput Intell Mag 15(1):64–75
Ba JL, Kiros JR, Hinton GE (2016) Layer normalization. arXiv preprint arXiv:1607.06450
Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align and translate. Technical report. arXiv preprint arXiv:1409.0473
Baziotis C, Pelekis N, Doulkeridis C (2017) Datastories at semeval-2017 task 4: deep LSTM with attention for message-level and topic-based sentiment analysis. In: Proceedings Of the 11th international workshop on semantic evaluation (SemEval-2017). Vancouver, Canada, August. Association For Computational Linguistics, pp 747–754
Bradbury J, Merity S, Xiong C, Socher R (2016) Quasi-recurrent neural networks. In Arxiv:1611.01576
Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140
Cambria E (2016) Affective computing and sentiment analysis. IEEE Intell Syst 31(2):102–107
Cambria K, Poria S, Gelbukh A, Thelwall M (2017) Sentiment analysis is a big suitcase. IEEE Intell Syst 32(6):74–80
Chaturvedi I, Ong YS, Tsang IW, Welsch RE, Cambria E (2016) Learning word dependencies in text by means of a deep recurrent belief network. Knowl Based Syst 108:144–154
Chaturvedi I, Cambria E, Welsch RE, Herrera F (2018) Distinguishing between facts and opinions for sentiment analysis: survey and challenges. Inf Fusion 44:65–77
Chen LC, Lee CM, Chen MY (2019a) Exploration of social media for sentiment analysis using deep learning. Springer, GmbH Germany, Part of Springer nature, soft computing
Cheng J, Zhao S, Zhang J, King I, Zhang X, Wang H (2017) Aspect-level sentiment classification with HEAT (HiErarchical ATtention) Network. CIKM 2017:97–106
Cheng J, Dong L, Lapata M (2016) Long short-term memory-networks for machine reading. ArXiv:1601.06733
Chen Z, Qian T (2019b) Transfer capsule network for aspect level sentiment classification. In: Proceedings of the 57th annual meeting of the association for computational linguistics, Florence, Italy, July 28–August 2, 2019. Association for Computational Linguistics, pp 547–556
Chen P, Sun Z, Bing L, Yang W (2017) Recurrent attention network on memory for aspect sentiment analysis. In: Proceedings of conference empirical methods natural language processing, pp 452–461
Cho K, Merrienboer BV, Bahdanau D, Bengio Y (2014) On the properties of neural machine translation: encoder-decoder approaches. Arxiv:1409.1259
Collobert RJ, Weston L, Bottou M, Karlen K, Kavukcuglu PK (2011) Natural language processing (almost) from scratch. J Mach Learn Res 12:2493–2537
Courbariaux M (2016) Binarized neural networks: training neural networks with weights and activations constrained to +1 or 1. Arxiv:1602.02830v3
Ding Y, Yu J, Jiang J (2017) Recurrent neural networks with auxiliary labels for cross-domain opinion target extraction. AAAI 2017:3436–3242
Dong L, Wei F, Tan C, Tang D, Zhou M, Xu K (2014) Adaptive recursive neural network for target-dependent twitter sentiment classification. ACL 2014:49–54
Dragoni M, Petrucci G (2017) A neural word embeddings approach for multi-domain sentiment analysis. IEEE TransAffect Comput 8(4):457–470
Dragoni M, Poria S, Cambria E (2018) OntoSenticNet: a commonsense ontology for sentiment analysis. IEEE Intell Syst 33(2):77–85
Du Y, Zhao X, He M, Guo W (2019a) A novel capsule based hybrid neural network for sentiment classification. IEEE Access 9:39321–39328
Duchi J, Hazan E, Singer Y (2011) Adaptive subgradient methods for online learning and stochastic optimization. J Mach Learn Res 12:21212159
Du C, Sun H, Wang J, Qi Q, Liao J, Xu T, Liu M (2019b) Capsule network with interactive attention for aspect-Level sentiment classification. In: Proceedings of international conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing, Hong Kong, China, November 3–7, 2019. Association for Computational Linguistics, pp 5489–5498
Du H, Xu X, Cheng X, Wu D, Liu Y, Yu Z (2016) Aspect-specific sentimental word embedding for sentiment analysis of online reviews. In: International conference companion on world wide web, pp 29–30
Ekbal A, Saha S (2011) Weighted vote-based classifier ensemble for named entity recognition: a genetic algorithm-based approach. ACM Trans Asian Lang Inf Process 10(2):1–9
Elman J (1990) Finding structure in time. Cognit Sci 14:179–211
Fan F, Feng Y, Zhao D (2018b) Multi-grained attention network for aspect-level sentiment classification. In: Proceedings of conference on empirical methods in natural language processing, pp 3433–3442
Fan C, Gao Q, Du J, Gui L, Xu R, Wong KF (2018a) Convolution-based memory network for aspect-based sentiment analysis. In: Proceedings of 41st International ACM SIGIR conference on Research and development in information retrieval, pp 1161–1164
Fentaw HW, Kim TH (2019) Design and investigation of capsule networks for sentence classification. Appl Sci 9:2200. https://doi.org/10.3390/app9112200
Fink CR, Chou DS, Kopecky JJ, Llorens AJ (2011) Coarse- and fine-grained sentiment analysis of social media text. Johns Hopkins Apl Tech Digest 30(1):22–30
Freund Y, Schapire RE (1996) Experiments with a new boosting algorithm. In: Proceedings of 13th international conference on machine learning (ICML), Bari, Italy, pp 148–156
Fu X, Liu W, Xu Y, Cui L (2017) Combine HowNet lexicon to train phrase recursive autoencoder for sentence-level sentiment analysis. Neurocomputing 241:18–27
Gao L, Guo Z, Zhang H, Xu X, Shen HT (2017) Video captioning with attention-based LSTM and semantic consistency. IEEE Trans Multimed 19(9):2045–2055
Graves A, Wayne G, Danihelka I (2014) Neural turing machines. ArXiv:1410.5401
Gu X, Gu Y, Wu H (2017) Cascaded convolutional neural networks for aspect-based opinion summary. Neural Process Lett 46(2):581–594
Guo B, Zhang C, Liu J, Ma X (2019) Improving text classification with weighted word embeddings via a multi-channel TextCNN model. Neurocomputing 363:366–374
Guo L, Ye H, Su W, Liu H, Sun K, Xiang H, (2018) Visualizing and understanding deep neural networks in CTR prediction. Sigir, 2018 Ecom, July 2018. Michigan, USA, Ann Arbor
Hassan A, Mahmood A (2018) Convolutional recurrent deep learning model for sentence classification. IEEE Access 6:13949–13957
Hassan J, Shoaib U (2019) Multi-class review rating classification using deep recurrent neural network. Springer, LLC, Part of Springer Nature, Neural Process Letters
He R, Lee WS, Ng HT, Dahlmeier D (2018) Exploiting document knowledge for aspect-level sentiment classification. In: ACL 2018
Hinton GE, Osindero S, Teh Y (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554
Huang EH, Socher R, Manning CD, Ng AY (2012) Improving word representations via global context and multiple word prototypes. In: ACL 2012
Huang B, Ou Y, Carley KM (2018) Aspect level sentiment classification with attention-over-attention neural networks. Proc SBP-BRiMS 2018:197–206
Irsoy O, Cardie C (2014a) Opinion mining with deep recurrent neural networks. In: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Doha, Qatar, October. Association For Computational Linguistics, pp 720–728
Irsoy O, Cardie C (2014b) Deep recursive neural networks for compositionality in language. In: Advances in neural information processing systems, pp. 2096–2104
Jabreel M, Moreno A (2017) Target-dependent sentiment analysis of tweets using a bi-directional gated recurrent unit. In: Proceedings of the 13th international conference on web information systems and technologies (WEBIST 2017), pp 80–87. ISBN: 978-989-758-246-2
Janocha K, Czarnecki WM (2017) On loss functions for deep neural networks in classification. Arxiv Preprint Arxiv:1702.05659
Jiang M, Liang Y, Feng X, Fan X, Pei Z, Xue Y, Guan R (2018) Text classification based on deep belief network and softmax regression. Neural Comput Appl 29(1):61–70. https://doi.org/10.1007/s00521-016-2401-x
Khan FH, Qamar U, Bashir S (2017) A semi-supervised approach to sentiment analysis using revised sentiment strength based on sentiwordnet. Knowl Inf Syst 51(3):851–872
Khanna, R, Awad, M (2015) Efficient learning machines: theories, concepts, and applications for engineers and system designers, Apress
Kim Y (2014) Convolutional neural networks for sentence classification. In: Proceedings of the 2014 conference on empirical methods. EMNLP, pp 1746–1751
Kingma DP, Lei BJ (2014) Adam: a method for stochastic optimization. Arxiv Preprint Arxiv:1412.6980
Krishnakumari K, Sivasankar E, Radhakrishnan S (2019) Hyperparameter tuning in convolutional neural networks for domain adaptation in sentiment classification. Soft Comput 24:3511–3527
Kumar A, Irsoy O, Su J, Bradbury J, English R, Pierce B, Ondruska P, Gulrajani I, Socher R (2016) Ask me anything: dynamic memory networks for natural language processing. In: ICML 2016
Lai S, Xu L, Liu K, Zhao J (2015) Recurrent convolutional neural networks for text classification. In: Twenty-ninth AAAI conference on artificial intelligence
Lakkaraju H, Socher R, Manning CD (2014) Aspect specific sentiment analysis using hierarchical deep learning. In: NIPS WS on deep neural networks and representation learning, pp 1–9
Lauren P, Qu G, Zhang F, Lendasse A (2018) Discriminant document embeddings with an extreme learning machine for classifying clinical narratives. Neurocomputing 277:129–138
Le QV, Ngiam J, Coates A, Lahiri A, Prochnow B, Ng AY (2011) On optimization methods for deep learning. In: ICML
Lebret R, Collobert R (2014) Word embeddings through Hellinger PCA. In: EACL 2014
LeCen Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436444
Li X, Bing L, Lam W, Shi B (2018) Transformation networks for target-oriented sentiment classification. ACL 2018:946–956
Liang X, Wang X, Lei Z, Liao S, Li SZ (2017) Soft-margin softmax for deep classification. In: Liu D, Xie S, Li Y, Zhao D, El-Alfy ES (eds) Neural information processing. ICONIP, 2017 Lecture notes in computer science, 10635. Springer, Cham
Lin Z, Feng M, dos Santos CN, Yu M, Xiang B, Zhou B, Bengio Y (2017) A structured self-attentive sentence embedding. arXiv preprint arXiv:1703.03130
Litjens G (2017) A survey on deep learning in medical image analysis. Med Image Anal 42:6088
Liu J, Zhang Y (2017a) Attention modeling for targeted sentiment. EACl 2017:572
Liu W, Wang Z, Liu X, Zeng N, Liu Y, Alsaadi FE (2017b) A survey of deep neural network architectures and their applications. Neurocomputing 234:1126
Liu N, Shen B, Zhang Z, Zhang Z, Mi K (2019) Attention-based Sentiment Reasoner for aspect-based sentiment analysis. Hum Cent Comput Inf Sci 9:35
Liu W, Cao G, Yin J (2019b) Bi-level attention model for sentiment analysis of short texts. IEEE Access 7:13–22
Liu S, Yang N, Li M, Zhou M (2014) A recursive recurrent neural network for statistical machine translation. In: ACL
Li H, Xu Z, Taylor G, Goldstein T (2017) Visualizing the loss landscape of neural nets. Arxiv Preprint Arxiv:1712.09913
Lo SL, Cambria E, Chiong R, Cornforth D (2017) Multilingual sentiment analysis: from formal to informal and scarce resource languages. Artif Intell Rev 48:499–527
Luong MT, Pham H, Manning CD (2015) Effective approaches to attention-based neural machine translation. arXiv:1508.04025
Ma S, Ji C (1998) A unified approach on fast training of feedforward and recurrent networks using EM algorithm. IEEE Trans Signal Process 46(8):2270–2274
Ma R, Wang K, Qiu T, Sangaiah AK, Lin D, Bin LH (2017) Feature-based compositing memory networks for aspect-based sentiment classification in social internet of things. Future Gener Comput Syst. https://doi.org/10.1016/j.future.2017.11.036
Ma Y, Peng H, Cambria E (2018) Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive LSTM. Proc AAAI 2018:5876–5883
Maas AL, Daly RE, Pham PT, Huang D, Ng AY, Potts C (2011) Learning word vectors for sentiment analysis. In: Proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies , vol 1, pp 142–150
Majumder N, Poria S, Gelbukh A, Akhtar MS, Cambria E, Ekbal A (2018) IARM: Inter-aspect relation modeling with memory networks in aspect-based sentiment analysis. Proc Conf Empir Methods Nat Lang Process 2018:3402–3411
Majumder N, Poria S, Peng H (2019) Sentiment and sarcasm classification with multitask learning. IEEE Intell Syst 34(3):38–43
Medhat W, Hassan A, Korashy H (2014) Sentiment analysis algorithms and applications: a survey. Ain Shams Eng J 5(4):1093–1113
Mikolov T, Sutskever I, Chen K, Corrado G, Dean J (2013b) Distributed representations of words and phrases and their compositionality. NIPS 2013:3111–3119
Mikolov T, Chen K, Corrado G, Dean J (2013a) Efficient estimation of word representations in vector space. In: ICLR 2013
Mnih A, Kavukcuoglu K (2013) Learning word embeddings efficiently with noise-contrastive estimation. In: NIPS 2013
Mousa A, Schuller S (2017) Contextual bidirectional long short-term memory recurrent neural network language models: a generative approach to sentiment analysis. In: Proceedings of the 15th conference of the European chapter of the association for computational linguistics, vol 1, pp 1023–1032
Nair V, Hinton GE (2010) Rectified linear units improve restricted boltzmann machines. In: ICML 2010
Nguyen HH, Yamagishi J, Echizen I (2019) Capsule-forensics: using capsule networks to detect forged images and videos. In: 2019 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 2307–2311
Nguyen N, Le A, Pham HT (2012) Deep bi-directional long short-term memory neural networks for sentiment analysis of social data. IUKM 2016:255–268
Paik I, Kwak T, Kim I (2019) Capsule networks need an improved routing algorithm. ArXiv, abs/1907.13327
Pandarachalil R, Sendhilkumar S, Mahalakshmi GS (2015) Twitter sentiment analysis for large-scale data: an unsupervised approach. Cognit Comput 7:254–262
Pascanu R, Gulcehre C, Cho K, Bengio Y (2014) How to construct deep recurrent neural networks. In: Proceedings of the second international conference on learning representations (ICLR 2014)
Pasupa K, Ayutthaya TSN (2019) Thai sentiment analysis with deep learning techniques: a comparative study based on word embedding, POS-tag, and sentic features. Sustain Cities Soc 50:101615
Patrick MK, Adekoya AF, Mighty AA, Edward BY (2019) Capsule networks-a survey. J King Saud Univ Comput Inf Sci (2019)
Pennington J, Socher R, Manning CD (2014) Glove: global vectors for word representation. In: Proceedings of the 2014 conference on empirical methods in natural language processing, pp 1532–1543
Pergola G, Gui L, He Y (2019) TDAM: a topic-dependent attention model for sentiment analysis. Inf Process Manag 56:102084
Plank B, Søgaard A, Goldberg Y (2016) Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss. In: Annual conference of the association for computational linguistics (ACL), pp 412–418
Pontiki M, Galanis D, Papageorgiou H, Manandhar S, Androutsopoulos I (2015) Semeval-2015 task 12: aspect based sentiment analysis. In: Proceedings of the 9th international workshop on semantic evaluation (SemEval 2015), pp 486–495
Pontiki M, Galanis D, Pavlopoulos J, Papageorgiou H, Androutsopoulos I, Manandhar S (2014) Semeval-2014 task 4: aspect based sentiment analysis. In: Proceedings of the 8th international workshop on semantic evaluation (SemEval 2014), pp 27–35
Qin P, Xu W, Guo J (2016) An empirical convolutional neural network approach for semantic relation classification. Neurocomputing 190:1–9
Rehman AU, Malik AK, Raza B, Ali W (2019) Hybrid CNN-LSTM model for improving accuracy of movie reviews sentiment analysis. Multimed Tools Appl. https://doi.org/10.1007/s11042-019-07788-7
Ren Y, Wang R, Ji D (2016) A topic-enhanced word embedding for Twitter sentiment classification. Inf Sci 369:188–198
Rezaeinia SM, Rahmani R, Ghodsi A, Veisi H (2018) Sentiment Analysis based on improved pre-trained word embeddings. Expert Syst Appl. https://doi.org/10.1016/j.eswa.2018.08.044
Rida-e-Fatima S, Javed A, Banjar A, Dawood AIH, Dawood H, Alamri A (2019) A multi-layer dual attention deep learning model with refined word embeddings for aspect-based sentiment analysis. IEEE Access 7:114795–114807
Rizk Y, Hajj N, Mitri N, Awad M (2019) Deep belief networks and cortical algorithms: a comparative study for supervised classification. Appl Comput Inf 15:81–93
Ruder S (2016) An overview of gradient descent optimization algorithms. Corr Abs/1609.04747
Sabour S, Frosst N, Hinton GE (2017) Dynamic routing between capsules. In: NIPS 2017
Sadr H, Pedram MM, Teshnehlab M (2019) A robust sentiment analysis method based on sequential combination of convolutional and recursive neural networks. Neural Process Lett. https://doi.org/10.1007/s11063-019-10049-1
Salehinejad H, Sankar S, Barfett J, Colak E, Valaee E (2017) Recent advances in recurrent neural networks. Arxiv Preprint Arxiv:1801.01078
Salinca A (2017) Convolutional neural networks for sentiment classification on business reviews. Arxiv Preprint Arxiv:1710.05978
Santos CD, Gatti G (2014) Deep convolutional neural networks for sentiment analysis of short texts. In: Proceedings of coling 2014, the 25th international conference on computational linguistics: technical papers, Dublin, Ireland, August 2014. Dublin City University And Association For Computational Linguistics, pp 69–78
Schmidhuber J (2015) Deep learning in neural networks: an overview. Neural Netw 61:85117
Schulz H, Behnke S (2012) Deep learning-layer-wise learning of feature hierarchies. Kunstliche Intelligenz 26(4):357–363
Schuster M, Paliwal KK (1997) Bidirectional recurrent neural networks. IEEE Trans Signal Process 45(11):2673–2681
Schuurmans J, Frasincar F (2019) Intent classification for dialogue utterances. IEEE Intell Syst. https://doi.org/10.1109/MIS.2019.2954966
Seifert C, Aamir A, Balagopalan A, Jain D, Sharma A, Grottel S, Gumhold S (2017) Visualizations of deep neural networks in computer vision: a survey. In: Transparent data mining for big and small data, pp 123–144
Severyn A, Moschitti A (2015) Twitter sentiment analysis with deep convolutional neural networks. In: Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval, pp 959–962
Shaw P, Uszkoreit J, Vaswani A (2018) Self-attention with relative position representations. In: Proceedings of NAACL
Shen T, Zhou T, Long G, Jiang J, Zhang C (2018) Bi-directional block self-attention for fast and memory-efficient sequence modeling. In: ICLR 2018
Socher R, Huval B, Manning CD, Ng AY (2012) Semantic compositionality through recursive matrix-vector spaces. EMNLP-CoNLL 12:1201–1211
Socher R, Perelygin A, Wu J, Chuang J, Manning C, Ng A, Potts C (2013) Recursive deep models for semantic compositionality over a sentiment treebank. In: EMNLP 2013
Srivastava RK, Greff K, Schmidhuber J (2015) Training very deep networks. Arxiv Preprint Arxiv:1507.06228
Sukhbaatar S, Szlam A, Weston J, Fergus R (2015) End-to-end memory networks. In: Cortes C, Lawrence ND, Lee DD, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates Inc, New York, pp 2440–2448
Sutskever I (2012) Training recurrent neural networks. Ph.D. Thesis, University of Toronto.
Tai KS, Socher R, Manning CD (2015) Improved semantic representations from tree-structured long short-term memory networks. In: ACL 2015
Tang D, Qin B, Feng X, Liu T (2015) Effective LSTMs for target dependent sentiment classification. Arxiv Preprint Arxiv:1512.01100
Tang D, Qin B, Liu T (2016) Aspect level sentiment classification with deep memory network. ArXiv Preprint ArXiv:1605:08900
Tang D, Wei F, Yang N, Zhou M, Liu T, Qin b (2014) Learning sentiment-specific word embedding for Twitter sentiment classification. In: Proceedings of the 52nd annual meeting of the association for computational linguistics, vol 1, pp 1555–1565
Tay Y, Tuan LA, Hui SC (2017) Dyadic memory networks for aspect-based sentiment analysis. CIKM 2017:107–116
Thireou T, Reczko M (2007) Bidirectional long short-term memory networks for predicting the subcellular localization of eukaryotic proteins. IEEE/ACM Trans Comput Biol Bioinform 4(3):441–446
Tian Z, Rong W, Shi L, Liu J, Xiong Z (2018) Attention aware bidirectional gated recurrent unit based framework for sentiment analysis. In: Liu W, Giunchiglia F, Yang B (eds) Knowledge science, engineering and management. Ksem, 2018. Lecture notes in computer science, 11061. Springer, Cham
Valdivia A, Luzón MV, Herrera F (2017) Sentiment analysis in tripAdvisor. IEEE Intell Syst 32(4):72–77
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017). Attention is all you need. In: ANIPS, pp 6000–6010
Wadawadagi RS, Pagi VB (2019a) An enterprise perspective of web content analysis research: a strategic road-map. Int J Knowl Web Intell 6(2):51–88
Wadawadagi RS, Pagi VB (2019b) A deep recursive neural network model for fine-grained opinion classification. In: Santosh K, Hegadi R (eds) Recent trends in image processing and pattern recognition. Rtip2r, 2018. Communications in computer and information science, 1037. Springer, Singapore
Wadawadagi RS, Pagi VB (2020a) Handbook of research on emerging trends and applications of machine learning. In: Arun S, Sandeep K, Anand N (eds) Chapter 24:Handbook of research on emerging trends and applications of machine learning. Part of the advances in computational intelligence and robotics book series. IGI-Gobal Publishers, Hershey, pp 508–527
Wadawadagi RS, Pagi VB (2020b) Fine-grained sentiment rating of online reviews with Deep-RNN. In: Niranjan NC, Takanori F (eds) Advances in artificial intelligence and data engineering. AIDE, 2019. Advances in intelligent systems and computing, 1133. Springer, Singapore
Wang X, Jiang W, Luo Z (2016) Combination of convolutional and recurrent neural network for sentiment analysis of short texts. In: Proceedings of the international conference on computational linguistics (COLING 2016)
Wang W, Pan SJ, Dahlmeier D (2017) Coupled multi-layer attentions for co-extraction of aspect and opinion terms. AAAI, pp 3316–3322. arxiv.org/abs/1702.01776
Wang W, Pan SJ, Dahlmeier D, Xiao X (2016a) Recursive neural conditional random fields for aspect-based sentiment analysis. In: EMNLP 2016
Wang Y, Sun A, Han J, Liu Y, Zhu X (2018) Sentiment analysis by capsules. In: Proceedings of the 2018 world wide web conference. International world wide web conferences steering committee, pp 1165–1174
Weichselbraun A, Gindl S, Fischer F, Vakulenko S, Scharl A (2017) Aspect-based extraction and analysis of affective knowledge from social media streams. IEEE Intell Syst 32(3):80–88
Weston J, Chopra S, Bordes A (2014) Memory networks. CoRR, abs/1410.3916
Wu H, Gu Y, Sun S, Gu X (2016) Aspect-based opinion summarization with convolutional neural networks. In: Proceedings of the international joint conference on neural networks, pp 3157–3163
Xiao T, Zhu J, Liu T (2013) Bagging and boosting statistical machine translation systems. Artif Intell 195:496–527
Xu C, Feng H, Yu G, Yang M, Wang X, Ao X (2020) Discovering protagonist of sentiment with aspect reconstructed capsule network. arXiv:1912.10785
Yin W, Kann K, Yu M, Schutze H (2017) Comparative study of CNN and RNN for natural language processing. Arxiv:1702.01923
Yosinski J, Clune J, Nguyen A, Fuchs T, Lipson H (2015) Understanding neural networks through deep visualization. In: ICML deep learning workshop (2015)
Yousefi-Azar M, Hamey L (2017) Text summarization using unsupervised deep learning. Expert Syst Appl 68:93–105
Yu LC, Wang J, Lai KR, Zhang X (2017) Refining word embeddings for sentiment analysis. In: Proceedings of the 2017 conference on empirical methods in natural language processing, pp 534–539 Copenhagen, Denmark, September 7–11, 2017
Yuan J, Zhao Y, Qin B, Liu T (2017) Local contexts are effective for neural aspect extraction. Commun Comput Inf Sci 774:244–255
Zeiler MD (2012) Adadelta: an adaptive learning rate method. Arxiv:1212.5701
Zhang L, Wang S, Liu B (2017) Deep learning for sentiment analysis: a survey. Wires Data Min Knowl Discov 2018:E1253
Zhang S, Zheng D, Hu X, Yang M (2015) Bidirectional long short-term memory networks for relation classification. In: Proceedings of the 29th Pacific Asia conference on language, information and computation, pp 73–78
Zhao W, Peng H, Eger S, Cambria E, Yang M (2019) Towards scalable and reliable capsule networks for challenging NLP applications. In:Proceedings of the 57th annual meeting of the association for computational linguistics, pp 1549–1559
Zhou S, Chen Q, Wang X (2014) Fuzzy deep belief networks for semi-supervised sentiment classification. Neurocomputing 131:312–322
Zhou X, Wan X, Xiao J (2016) Attention-based LSTM network for cross-lingual sentiment classification. In: EMNLP 2016
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wadawadagi, R., Pagi, V. Sentiment analysis with deep neural networks: comparative study and performance assessment. Artif Intell Rev 53, 6155–6195 (2020). https://doi.org/10.1007/s10462-020-09845-2
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10462-020-09845-2