Skip to main content

Advertisement

Log in

BARLAT: A Nearly Unsupervised Approach for Aspect Category Detection

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

Aspect category detection is an essential task in aspect-based sentiment analysis. Most previous works use labeled data and apply a supervised learning approach to detect the aspect category. However, to avoid the dependency on labeled data, some researchers have also applied unsupervised learning approaches, wherein variants of topic models and neural network-based models have been built for this task. These unsupervised methods focus on co-occurrences of words and ignore the contextual meaning of the words in the given sentence. Thus, such models perform reasonably well in detecting the explicitly expressed aspect category but often fail in identifying the implicitly expressed aspect category in the sentence. This paper focuses on the contextual meaning of the word. It adopts a document clustering approach requiring minimal user guidance, i.e., only a small set of seed words for each aspect category to detect efficiently implicit and explicit aspect categories. A novel BERT-based Attentive Representation Learning with Adversarial Training (BARLAT) model is presented in this paper, which utilizes domain-based contextual word embedding (BERT) for generating the sentence representation and uses these representations for clustering the sentences through attentive representation learning. Further, the model parameters are generalized better by performing adversarial training, which adds perturbations to the cluster representations. BARLAT is the first nearly unsupervised method that uses the contextual meaning of the words for learning the aspect categories through an adversarial attentive learning approach. The performance of BARLAT is compared with various state-of-the-art models using F1-score on Laptop and Restaurant datasets. The experimental results show that BARLAT outperforms the best existing model by a margin of 1.1% and 2.3% on Restaurant and Laptop datasets, respectively.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Notes

  1. http://www.cs.cmu.edu/~mehrbod/RR/.

References

  1. Hu M, Liu B (2004) Mining and summarizing customer reviews. In: Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining, pp 168–177

  2. Liu B, Zhang L (2012) A survey of opinion mining and sentiment analysis. In: Mining text data. Springer, pp 415–463

  3. Pontiki M, Galanis D, Papageorgiou H, Androutsopoulos I, Manandhar S, Al-Smadi M, Al-Ayyoub M, Zhao Y, Qin B, De Clercq O, et al. (2016)Semeval-2016 task 5: Aspect based sentiment analysis. In: 10th international workshop on semantic evaluation (SemEval 2016)

  4. Qiu G, Liu B, Jiajun B, Chen C (2011) Opinion word expansion and target extraction through double propagation. Comput Linguist 37(1):9–27

    Article  Google Scholar 

  5. Blei DM, Ng AY, Jordan MI (2003) Latent dirichlet allocation. J Mach Learn Res 3:993–1022

    MATH  Google Scholar 

  6. Brody S, Elhadad N (201) An unsupervised aspect-sentiment model for online reviews. In: Human language technologies: the 2010 annual conference of the North American chapter of the association for computational linguistics. Association for Computational Linguistics, pp 804–812

  7. Zhao WX, Jiang J, Yan H, Li X (2010) Jointly modeling aspects and opinions with a maxent-lda hybrid. In: Proceedings of the 2010 conference on empirical methods in natural language processing. Association for Computational Linguistics, pp 56–65

  8. Chen Z, Mukherjee A, Liu B (2014) Aspect extraction with automated prior knowledge learning. In: Proceedings of the 52nd annual meeting of the association for computational linguistics (volume 1: long papers), pp 347–358

  9. Mimno D, Wallach H, Talley E, Leenders M, McCallum A (2011) Optimizing semantic coherence in topic models. In: Proceedings of the 2011 conference on empirical methods in natural language processing, pp 262–272

  10. Miao Y, Yu L, Blunsom P (2016) Neural variational inference for text processing. In: International conference on machine learning, pp 1727–1736

  11. Srivastava A, Sutton C (2017) Autoencoding variational inference for topic models. Preprint arXiv:1703.01488

  12. Liao M, Li J, Zhang H, Wang L, Wu X, Wong K-F (2019) Coupling global and local context for unsupervised aspect extraction. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pp 4571–4581

  13. He R, Lee WS, Ng HT, Dahlmeier D (2017) An unsupervised neural attention model for aspect extraction. In: Proceedings of the 55th annual meeting of the association for computational linguistics (volume 1: long papers), pp 388–397

  14. Tulkens S, van Cranenburgh A (2020) Embarrassingly simple unsupervised aspect extraction. Preprint arXiv:2004.13580

  15. Mikolov T, Chen K, Corrado G, Dean J (2013) Efficient estimation of word representations in vector space. Preprint arXiv:1301.3781

  16. Devlin J, Chang M-W, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. Preprint arXiv:1810.04805

  17. Titov I, McDonald R (2008) Modeling online reviews with multi-grain topic models. In: Proceedings of the 17th international conference on World Wide Web, pp 111–120

  18. Mukherjee A, Liu B (2012) Aspect extraction through semi-supervised modeling. In: Proceedings of the 50th annual meeting of the association for computational linguistics: Long papers-volume 1. Association for Computational Linguistics, pp 339–348

  19. Yan X, Guo J, Lan Y, Cheng X (2013) A biterm topic model for short texts. In: Proceedings of the 22nd international conference on World Wide Web, pp 1445–1456

  20. Zbontar J, Jing L, Misra I, LeCun Y, Deny S (2021) Barlow twins: self-supervised learning via redundancy reduction

  21. Lara JS, González FA (2020) Dissimilarity mixture autoencoder for deep clustering. Preprint arXiv:2006.08177

  22. Hadifar A, Sterckx L, Demeester T, Develder C (2019) A self-training approach for short text clustering. In: Proceedings of the 4th workshop on representation learning for NLP (RepL4NLP-2019), pp 194–199

  23. Zhang W, Dong C, Yin J, Wang J (2019)Attentive representation learning with adversarial training for short text clustering. Preprint arXiv:1912.03720

  24. García-Pablos A, Cuadros M, Rigau G (2018) W2vlda: almost unsupervised system for aspect based sentiment analysis. Exp Syst Appl 91:127–137

    Article  Google Scholar 

  25. Angelidis S, Lapata M (2018) Summarizing opinions: aspect extraction meets sentiment prediction and they are both weakly supervised. Preprint arXiv:1808.08858

  26. Sokhin T, Khodorchenko M, Butakov N (2020) Unsupervised neural aspect search with related terms extraction. Preprint arXiv:2005.02771

  27. Szegedy C, Zaremba W, Sutskever I, Bruna J, Erhan D, Goodfellow I, Fergus R (2013) Intriguing properties of neural networks. Preprint arXiv:1312.6199

  28. Goodfellow IJ, Shlens J, Szegedy C (2014) Explaining and harnessing adversarial examples. Preprint arXiv:1412.6572

  29. Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, Krikun M, Cao Y, Gao Q, Macherey K, et al. (2016) Google’s neural machine translation system: bridging the gap between human and machine translation. Preprint arXiv:1609.08144

  30. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems, pp 5998–6008

  31. Xiao H (2018) bert-as-service. https://github.com/hanxiao/bert-as-service

  32. Socher R, Karpathy A, Le QV, Manning CD, Ng AY (2014) Grounded compositional semantics for finding and describing images with sentences. Trans Assoc Comput Linguist 2:207–218

    Article  Google Scholar 

  33. Weston J, Bengio S, Usunier N (2011) Wsabie: scaling up to large vocabulary image annotation. In: Proceedings of the twenty-second international joint conference on artificial intelligence—volume volume three, IJCAI’11. AAAI Press, pp 2764–2770

  34. Iyyer M, Guha A, Chaturvedi S, Boyd-Graber J, Daumé III H (2016) Feuding families and former Friends: unsupervised learning for dynamic fictional relationships. In: Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies, San Diego, California, June . Association for Computational Linguistics, pp 1534–1544

  35. Miyato T, Maeda S-i, Koyama M, Nakae K, Ishii S (2015) Distributional smoothing with virtual adversarial training. Preprint arXiv:1507.00677

  36. Wang L, Liu K, Cao Z, Zhao J, De Melo G (2015)Sentiment-aspect extraction based on restricted boltzmann machines. In: Proceedings of the 53rd annual meeting of the association for computational linguistics and the 7th international joint conference on natural language processing (volume 1: long papers), pp 616–625

  37. He R, McAuley J (2016) Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In: proceedings of the 25th international conference on world wide web, pp 507–517

  38. Li N, Chow C-Y, Zhang J-D (2019) Seeded-btm: enabling biterm topic model with seeds for product aspect mining. In: 2019 IEEE 21st international conference on high performance computing and communications; IEEE 17th international conference on smart city; IEEE 5th international conference on data science and systems (HPCC/SmartCity/DSS). IEEE, pp 2751–2758

  39. Goldberg Y, Levy O (2014) word2vec explained: deriving mikolov et al.’s negative-sampling word-embedding method. Preprint arXiv:1402.3722

  40. Vargas DS, Pessutto LRC, Moreira VP (2020)Simple unsupervised similarity-based aspect extraction. Preprint arXiv:2008.10820

  41. Luo L, Ao X, Song Y, Li J, Yang X, He Q, Yu D (2019)Unsupervised neural aspect extraction with sememes. In: IJCAI, pp 5123–5129

  42. Xu H, Liu B, Shu L, Yu PS (2019) Bert post-training for review reading comprehension and aspect-based sentiment analysis. Preprint arXiv:1904.02232

  43. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. Preprint arXiv:1412.6980

  44. Conneau A, Khandelwal K, Goyal N, Chaudhary V, Wenzek G, Guzmán F, Grave E, Ott M, Zettlemoyer L, Stoyanov V (2019) Unsupervised cross-lingual representation learning at scale. Preprint arXiv:1911.02116

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Avinash Kumar.

Ethics declarations

Conflict of interest

The authors have no relevant financial or non-financial interests to disclose. The authors have no conflicts of interest to declare that are relevant to the content of this article. All authors certify that they have no affiliations with or involvement in any organization or entity with any financial interest or non-financial interest in the subject matter or materials discussed in this manuscript. The authors have no financial or proprietary interests in any material discussed in this article.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kumar, A., Gupta, P., Kotak, N. et al. BARLAT: A Nearly Unsupervised Approach for Aspect Category Detection. Neural Process Lett 54, 4495–4519 (2022). https://doi.org/10.1007/s11063-022-10819-4

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-022-10819-4

Keywords

Navigation