skip to main content
research-article

Self-Adaptive Representation Learning Model for Multi-Modal Sentiment and Sarcasm Joint Analysis

Authors Info & Claims
Published:11 January 2024Publication History
Skip Abstract Section

Abstract

Sentiment and sarcasm are intimate and complex, as sarcasm often deliberately elicits an emotional response in order to achieve its specific purpose. Current challenges in multi-modal sentiment and sarcasm joint detection mainly include multi-modal representation fusion and the modeling of the intrinsic relationship between sentiment and sarcasm. To address these challenges, we propose a single-input stream self-adaptive representation learning model (SRLM) for sentiment and sarcasm joint recognition. Specifically, we divide the image into blocks to learn its serialized features and fuse textual feature as input to the target model. Then, we introduce an adaptive representation learning network using a gated network approach for sarcasm and sentiment classification. In this framework, each task is equipped with its dedicated expert network responsible for learning task-specific information, while the shared expert knowledge is acquired and weighted through the gating network. Finally, comprehensive experiments conducted on two publicly available datasets, namely Memotion and MUStARD, demonstrate the effectiveness of the proposed model when compared to state-of-the-art baselines. The results reveal a notable improvement on the performance of sentiment and sarcasm tasks.

REFERENCES

  1. [1] Farias DI Hernández and Rosso Paolo. 2017. Irony, sarcasm, and sentiment analysis. Sentiment Analysis in Social Networks. Elsevier, 113128.Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Saddik Abdulmotaleb El, Fischer Stefan, and Steinmetz Ralf. 2001. Reusable multimedia content in Web based learning systems. IEEE MultiMedia 8, 3 (2001), 3038.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. [3] Moin Anam, Aadil Farhan, Ali Zeeshan, and Kang Dongwann. 2023. Emotion recognition framework using multiple modalities for an effective human–computer interaction. The Journal of Supercomputing 79, 8 (2023), 93209349.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. [4] Zhang Xiaoheng and Li Yang. 2023. A cross-modality context fusion and semantic refinement network for emotion recognition in conversation. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 1309913110.Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Hossain M. Shamim and Muhammad Ghulam. 2019. Emotion recognition using deep learning approach from audio-visual emotional big data. Information Fusion 49 (2019), 6978.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. [6] Yue Deng, Wenxuan Zhang, Sinno Jialin Pan, and Lidong Bing. 2023. Bidirectional generative framework for cross-domain aspect-based sentiment analysis. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics. 12272–12285.Google ScholarGoogle Scholar
  7. [7] Liu Yi, Zheng Zengwei, Zhou Binbin, Ma Jianhua, Sun Lin, and Xia Ruichen. 2022. Multimodal sarcasm detection based on multimodal sentiment co-training. In Proceedings of the 2022 IEEE Smartworld, Ubiquitous Intelligence & Computing, Scalable Computing & Communications, Digital Twin, Privacy Computing, Metaverse, Autonomous & Trusted Vehicles (SmartWorld/UIC/ScalCom/DigitalTwin/PriComp/Meta). IEEE, 508515.Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Wen Changsong, Jia Guoli, and Yang Jufeng. 2023. DIP: Dual incongruity perceiving network for sarcasm detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 25402550.Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Riloff Ellen, Qadir Ashequl, Surve Prafulla, Silva Lalindra De, Gilbert Nathan, and Huang Ruihong. 2013. Sarcasm as contrast between a positive sentiment and negative situation. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing. 704714.Google ScholarGoogle Scholar
  10. [10] Zhao Haojie, Wang Xiao, Wang Dong, Lu Huchuan, and Ruan Xiang. 2023. Transformer vision-language tracking via proxy token guided cross-modal fusion. Pattern Recognition Letters 168 (2023), 1016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. [11] Yazhou Zhang, Yang Yu, Dongming Zhao, Zuhe Li, Bo Wang, Yuexian Hou, Prayag Tiwari, and Jing Qin. 2023. Learning multi-task commonness and uniqueness for multi-modal sarcasm detection and sentiment analysis in conversation. IEEE Transactions on Artificial Intelligence, 1, 1 (2023), 1–13.Google ScholarGoogle Scholar
  12. [12] Md. Shad Akhtar, Dushyant Singh Chauhan, Deepanway Ghosal, Soujanya Poria, Asif Ekbal, and Pushpak Bhat-tacharyya. 2019. Multi-task learning for multi-modal emotion recognition and sentiment analysis. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT. 370–379.Google ScholarGoogle Scholar
  13. [13] Yang Bo, Wu Lijun, Zhu Jinhua, Shao Bo, Lin Xiaola, and Liu Tie-Yan. 2022. Multimodal sentiment analysis with two-phase multi-task learning. IEEE/ACM Transactions on Audio, Speech, and Language Processing 30 (2022), 20152024.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An image is worth \(16\times 16\) words: Transformers for image recognition at scale. In 9th International Conference on Learning Representations (ICLR’21).Google ScholarGoogle Scholar
  15. [15] Albraikan Amani, Tobón Diana P., and Saddik Abdulmotaleb El. 2018. Toward user-independent emotion recognition using physiological signals. IEEE sensors Journal 19, 19 (2018), 84028412.Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Albraikan Amani, Hafidh Basim, and Saddik Abdulmotaleb El. 2018. iAware: A real-time emotional biofeedback system based on physiological signals. IEEE Access 6 (2018), 7878078789.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Vaswani Ashish, Shazeer Noam, Parmar Niki, Uszkoreit Jakob, Jones Llion, Gomez Aidan N, Kaiser Łukasz, and Polosukhin Illia. 2017. Attention is all you need. In Proceedings of the 31st International Conference Advances in Neural Information Processing Systems 30 (2017).Google ScholarGoogle Scholar
  18. [18] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT. 4171–4186.Google ScholarGoogle Scholar
  19. [19] He Kaiming, Zhang Xiangyu, Ren Shaoqing, and Sun Jian. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 770778.Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] He Jiaxuan and Hu Haifeng. 2021. MF-BERT: Multimodal fusion in pre-trained BERT for sentiment analysis. IEEE Signal Processing Letters 29 (2021), 454458.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Kumar Ayush and Vepa Jithendra. 2020. Gated mechanism for attention based multi modal sentiment analysis. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing. 44774481. DOI:Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Ghosal Deepanway, Akhtar Md Shad, Chauhan Dushyant, Poria Soujanya, Ekbal Asif, and Bhattacharyya Pushpak. 2018. Contextual inter-modal attention for multi-modal sentiment analysis. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 34543466.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Zhang Qiongan, Shi Lei, Liu Peiyu, Zhu Zhenfang, and Xu Liancheng. 2023. ICDN: Integrating consistency and difference networks by transformer for multimodal sentiment analysis. Applied Intelligence 53, 12 (2023), 1633216345.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. [24] Peng Junjie, Wu Ting, Zhang Wenqiang, Cheng Feng, Tan Shuhua, Yi Fen, and Huang Yansong. 2023. A fine-grained modal label-based multi-stage network for multimodal sentiment analysis. Expert Systems with Applications 221 (2023), 119721.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. [25] Kim Kyeonghun and Park Sanghyun. 2023. AOBERT: All-modalities-in-One BERT for multimodal sentiment analysis. Information Fusion 92 (2023), 3745.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Schifanella Rossano, Juan Paloma De, Tetreault Joel, and Cao Liangliang. 2016. Detecting sarcasm in multimodal social platforms. In Proceedings of the 24th ACM International Conference on Multimedia. 11361145.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. [27] Wang Xinyu, Sun Xiaowen, Yang Tan, and Wang Hongbo. 2020. Building a bridge: A method for image-text sarcasm detection without pretraining on image-text data. In Proceedings of the 1st International Workshop on Natural Language Processing Beyond Text. 1929.Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Ding Ning, Tian Sheng-wei, and Yu Long. 2022. A multimodal fusion method for sarcasm detection based on late fusion. Multimedia Tools and Applications 81, 6 (2022), 85978616.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. [29] Cai Yitao, Cai Huiyu, and Wan Xiaojun. 2019. Multi-modal sarcasm detection in twitter with hierarchical fusion model. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 25062515.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Qiao Yang, Jing Liqiang, Song Xuemeng, Chen Xiaolin, Zhu Lei, and Nie Liqiang. 2023. Mutual-enhanced incongruity learning network for multi-modal sarcasm detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 37. 95079515.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. [31] Lu Xinkai, Qian Ying, Yang Yan, and Pang Wenrao. 2022. Sarcasm detection of dual multimodal contrastive attention networks. In Proceedings of the 2022 IEEE Smartworld, Ubiquitous Intelligence & Computing, Scalable Computing & Communications, Digital Twin, Privacy Computing, Metaverse, Autonomous & Trusted Vehicles (SmartWorld/UIC/ScalCom/DigitalTwin/PriComp/Meta). IEEE, 14551460.Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Liang Bin, Lou Chenwei, Li Xiang, Yang Min, Gui Lin, He Yulan, Pei Wenjie, and Xu Ruifeng. 2022. Multi-modal sarcasm detection via cross-modal graph convolutional network. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vol. 1. Association for Computational Linguistics, 17671777.Google ScholarGoogle ScholarCross RefCross Ref
  33. [33] Hui Liu, Wenya Wang, and Haoliang Li. 2022. Towards multi-modal sarcasm detection via hierarchical congruity modeling with knowledge enhancement. In 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP’22). 4995–5006.Google ScholarGoogle Scholar
  34. [34] Tan Yue, Wang Bo, Liu Anqi, Zhao Dongming, Huang Kun, He Ruifang, and Hou Yuexian. 2023. Guiding dialogue agents to complex semantic targets by dynamically completing knowledge graph. In Findings of the Association for Computational Linguistics: ACL 2023. 65066518.Google ScholarGoogle Scholar
  35. [35] Majumder Navonil, Poria Soujanya, Peng Haiyun, Chhaya Niyati, Cambria Erik, and Gelbukh Alexander. 2019. Sentiment and sarcasm classification with multitask learning. IEEE Intelligent Systems 34, 3 (2019), 3843.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Vitman Oxana, Kostiuk Yevhen, Sidorov Grigori, and Gelbukh Alexander. 2023. Sarcasm detection framework using context, emotion and sentiment features. Expert Systems with Applications 234 (2023), 121068.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Yin Chunyan, Chen Yongheng, and Zuo Wanli. 2021. Multi-task deep neural networks for joint sarcasm detection and sentiment analysis. Pattern Recognition and Image Analysis 31 (2021), 103108.Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. [38] Chauhan Dushyant Singh, Dhanush S.R., Ekbal Asif, and Bhattacharyya Pushpak. 2020. Sentiment and emotion help sarcasm? A multi-task learning framework for multi-modal sarcasm, sentiment and emotion analysis. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 43514360.Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Yaochen Liu, Yazhou Zhang, and Dawei Song. 2023. A quantum probability driven framework for joint multi-modal sarcasm, sentiment and emotion analysis. IEEE Transactions on Affective Computing 1 (2023), 1–15.Google ScholarGoogle Scholar
  40. [40] Hossain M. Shamim and Muhammad Ghulam. 2018. Emotion-aware connected healthcare big data towards 5g. IEEE Internet of Things Journal 5, 4 (2018), 23992406.Google ScholarGoogle ScholarCross RefCross Ref
  41. [41] Sharma Chhavi, Paka William, Scott Deepesh Bhageria, Das Amitava, Poria Soujanya, Chakraborty Tanmoy, and Gambäck Björn. 2020. Task report: Memotion analysis 1.0@ semeval 2020: The visuo-lingual metaphor. In Proceedings of the 14th International Workshop on Semantic Evaluation, Sep. Association for Computational Linguistics.Google ScholarGoogle Scholar
  42. [42] Cramer Aurora Linh, Wu Ho-Hsiang, Salamon Justin, and Bello Juan Pablo. 2019. Look, listen, and learn more: Design choices for deep audio embeddings. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 38523856.Google ScholarGoogle ScholarCross RefCross Ref
  43. [43] Potamias Rolandos Alexandros, Siolas Georgios, and Stafylopatis Andreas-Georgios. 2020. A transformer-based approach to irony and sarcasm detection. Neural Computing and Applications 32 (2020), 1730917320.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. [44] Tan Mingxing and Le Quoc. 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning. PMLR, 61056114.Google ScholarGoogle Scholar
  45. [45] Amir Zadeh, Minghai Chen, Soujanya Poria, Erik Cambria, and Louis-Philippe Morency. 2017. Tensor fusion network for multimodal sentiment analysis. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. 1103–1114.Google ScholarGoogle Scholar
  46. [46] Pramanick Shraman, Roy Aniket, and Patel Vishal M.. 2022. Multimodal learning using optimal transport for sarcasm and humor detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 39303940.Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Tsai Yao-Hung Hubert, Bai Shaojie, Liang Paul Pu, Kolter J. Zico, Morency Louis-Philippe, and Salakhutdinov Ruslan. 2019. Multimodal transformer for unaligned multimodal language sequences. In Proceedings of the Conference Association for Computational Linguistics. Meeting, Vol. 2019. NIH Public Access, 6558.Google ScholarGoogle ScholarCross RefCross Ref
  48. [48] Wei Yinwei, Wang Xiang, Nie Liqiang, He Xiangnan, Hong Richang, and Chua Tat-Seng. 2019. MMGCN: Multi-modal graph convolution network for personalized recommendation of micro-video. In Proceedings of the 27th ACM International Conference on Multimedia. 14371445.Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. [49] Yazhou Zhang, Ao Jia, Bo Wang, Peng Zhang, Dongming Zhao, Pu Li, Yuexian Hou, Xiaojia Jin, Dawei Song, and Jing Qin. 2023. M3gat: A multi-modal, multi-task interactive graph attention network for conversational sentiment analysis and emotion recognition. ACM Transactions on Information Systems, 42, 1 (2023), 1–32.Google ScholarGoogle Scholar
  50. [50] Chaturvedi Iti, Su Chit Lin, and Welsch Roy E.. 2021. Fuzzy aggregated topology evolution for cognitive multi-tasks. Cognitive Computation 13 (2021), 96107.Google ScholarGoogle ScholarCross RefCross Ref
  51. [51] George-Alexandru Vlad, George-Eduard Zaharia, Dumitru-Clementin Cercel, Costin Chiru, and Stefan Trausan Matu. 2020. Upb at semeval-2020 task 8: Joint textual and visual modeling in a multi-task learning architecture for memotion analysis. In Proceedings of the Fourteenth Workshop on Semantic Evaluation. 1208–1214.Google ScholarGoogle Scholar

Index Terms

  1. Self-Adaptive Representation Learning Model for Multi-Modal Sentiment and Sarcasm Joint Analysis

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in

        Full Access

        • Published in

          cover image ACM Transactions on Multimedia Computing, Communications, and Applications
          ACM Transactions on Multimedia Computing, Communications, and Applications  Volume 20, Issue 5
          May 2024
          650 pages
          ISSN:1551-6857
          EISSN:1551-6865
          DOI:10.1145/3613634
          • Editor:
          • Abdulmotaleb El Saddik
          Issue’s Table of Contents

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 11 January 2024
          • Online AM: 5 December 2023
          • Accepted: 27 November 2023
          • Revised: 23 November 2023
          • Received: 27 September 2023
          Published in tomm Volume 20, Issue 5

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
        • Article Metrics

          • Downloads (Last 12 months)285
          • Downloads (Last 6 weeks)71

          Other Metrics

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        Full Text

        View this article in Full Text.

        View Full Text