ABSTRACT
This paper describes the method that we submitted to the FinSim-2 task on learning similarities for the financial domain. This task aims to automatically classify the Financial domain terms into the most relevant hypernym (or top-level) concept in an external ontology. This paper shows the result of experiments using the Catboost, Attention-LSTM, BERT, RoBERTa to develop an automatic finance domain classifier via word ontology and embedding. The experiment result demonstrates that each model could be an effective method to tackle the FinSim-2 task, respectively.
- Colin Raffel and Daniel P. W. Ellis. 2015. Feed-forward networks with attention can solve some long term memory problems. https://arxiv.org/abs/1512.08756.Google Scholar
- Jacob Devlin, Ming-Wei Chang, Ken ton Lee, and Kristina Toutanova.2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.Google Scholar
- Dogu Araci. 2019. FinBERT: Financial Sentiment Analysis with Pre-trained Language Models. https://arxiv.org/abs/1908.10063.Google Scholar
- Ke Tian, Hua Chen and Jie Yang. 2020. aiai at FinSBD task: Sentence Boundary Detection in Noisy Texts From Financial Documents Using Deep Attention Model. Processing of the First Workshop on Finance Technology and Natural Language Processing IJCAI2020. Japan.Google Scholar
- Ke Tian and Zi Jun Peng. 2019. aiai at FinNum task: Financial numeral tweets fine-grained classification using deep word and character embed- ding-based attention model. The 14th NTCIR Conference, Tokyo, Japan.Google Scholar
- FinSim-2 task. 2021. https://sites.google.com/nlg.csie.ntu.edu.tw/finweb2021/shared-task-finsim-2.Google Scholar
- Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg Corrado and Jeffrey Dean. 2013. Distributed representations of words and phrases and their composi- tionality. https://arxiv.org/abs/1310.4546.Google Scholar
- Anna Veronika Dorogush, Vasily Ershov, Andrey Gulin. 2018. CatBoost: gradient boosting with categorical features support. https://arxiv.org/abs/1810.11363.Google Scholar
- Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. https://arxiv.org/abs/1907.11692.Google Scholar
- Natural Language Toolkit. https://www.nltk.org.Google Scholar
- aiai at the FinSim-2 task: Finance Domain Terms Automatic Classification Via Word Ontology and Embedding
Recommendations
GOAT at the FinSim-2 task: Learning Word Representations of Financial Data with Customized Corpus
WWW '21: Companion Proceedings of the Web Conference 2021In this paper, we present our approaches for the FinSim 2021 Shared Task on Learning Semantic Similarities for the Financial Domain. The aim of the FinSim shared task is to automatically classify a given list of terms from the financial domain into the ...
TCS_WITM_2021 @FinSim-2: Transformer based Models for Automatic Classification of Financial Terms
WWW '21: Companion Proceedings of the Web Conference 2021Recent advancement in neural network architectures has provided several opportunities to develop systems to automatically extract and represent information from domain specific unstructured text sources. The Finsim-2021 shared task, collocated with the ...
JSI at the FinSim-2 task: Ontology-Augmented Financial Concept Classification
WWW '21: Companion Proceedings of the Web Conference 2021Ontologies are increasingly used for machine reasoning over the last few years. They can provide explanations of concepts or be used for concept classification if there exists a mapping from the desired labels to the relevant ontology. Another ...
Comments