Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3543106.3543120acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicemcConference Proceedingsconference-collections
research-article
Open access

Extract Aspect-based Financial Opinion Using Natural Language Inference

Published: 22 August 2022 Publication History

Abstract

The emergence of transformer-based pre-trained language models (PTLMs) has bought new and improved techniques to natural language processing (NLP). Traditional rule-based NLP, for instance, is known for its deficiency of creating context-aware representations of words and sentences. Natural language inference (NLI) addresses this deficiency by using PTLMs to create context-sensitive embedding for contextual reasoning. This paper outlines a system design that uses traditional rule-based NLP and deep learning to extract aspect-based financial opinion from financial commentaries written using colloquial Cantonese, a dialect of the Chinese language used in Hong Kong. We need to confront the issue that existing off-the-shelf PTLMs, such as BERT and Roberta, are not pre-trained to understand the language semantics of colloquial Cantonese, let alone the slang, jargon, and codeword that people in Hong Kong use to articulate opinions. As a result, we approached the opinion extraction problem differently from the mainstream approaches, which use model-based named entity recognition (NER) to detect and extract opinion aspects as named entities and named entity relations. Because there is no PTLM for our specific language and problem domain, we solve the opinion extraction problem using rule-based NLP and deep learning techniques. We report our experience of creating a lexicon and identifying candidate opinion aspects in the input text using rule-based NLP. We discuss how to improve BERT’s linguistic knowledge of colloquial Cantonese through a fine-tuning procedure. We illustrate how to prepare the input text for contextual reasoning and demonstrate how to use NLI to confirm candidate opinion aspects as extractable.

References

[1]
Chung-Chi Chen, Hen-Hsen Huang, and Hsin-Hsi Chen. 2020. Fine-grained Financial Opinion Mining: A Survey and Research Agenda. https://doi.org/10.48550/ARXIV.2005.01897
[2]
Chung-Chi Chen, Hen-Hsen Huang, and Hsin-Hsi Chen. 2021. From Opinion Mining to Financial Argument Mining.
[3]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, 4171–4186. https://doi.org/10.18653/v1/N19-1423
[4]
T. Gao, A. Fisch, and D. Chen. 2021. Making pre-trained language models better few-shot learners.
[5]
Suchin Gururangan, Ana Marasović, Swabha Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, and Noah A. Smith. 2020. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks. https://doi.org/10.48550/ARXIV.2004.10964
[6]
Wenfei Hu, Lu Liu, Yupeng Sun, Yu Wu, Zhicheng Liu, Ruixin Zhang, and Tao Peng. 2022. NLIRE: A Natural Language Inference method for Relation Extraction. Journal of Web Semantics 72 (2022). https://doi.org/10.1016/j.websem.2021.100686
[7]
Avinash Kumar, Aditya Srikanth Veerubhotla, Vishnu Teja Narapareddy, Vamshi Aruru, Lalita Bhanu Murthy Neti, and Aruna Malapati. 2021. Aspect term extraction for opinion mining using a Hierarchical Self-Attention Network. Neurocomputing 465(2021). https://doi.org/10.1016/j.neucom.2021.08.133
[8]
Macedo Maia, Siegfried Handschuh, André Freitas, Brian Davis, Ross McDermott, Manel Zarrouk, and Alexandra Balahur. 2018. WWW’18 Open Challenge: Financial Opinion Mining and Question Answering. In Proceedings of the Web Conference 2018.
[9]
Xipeng Qiu, TianXiang Sun, Yige Xu, Yunfan Shao, Ning Dai, and Xuanjing Huang. 2020. Pre-trained Models for Natural Language Processing: A Survey. SCIENCE CHINA Technological Sciences 63, 10 (2020), 1872–1897. https://doi.org/10.1007/s11431-020-1647-3
[10]
Yi Sun, Yu Zheng, Chao Hao, and Hangping Qiu. 2021. NSP-BERT: A Prompt-based Zero-Shot Learner Through an Original Pre-training Task-Next Sentence Prediction. ArXiv (2021).
[11]
Sinong Wang, Han Fang, Madian Khabsa, Hanzi Mao, and Hao Ma. 2021. Entailment as Few-Shot Learner. https://doi.org/10.48550/ARXIV.2104.14690
[12]
Zexuan Zhong and Danqi Chen. 2020. A Frustratingly Easy Approach for Entity and Relation Extraction. https://doi.org/10.48550/ARXIV.2010.12812

Cited By

View all
  • (2024)Credibility-based knowledge graph embedding for identifying social brand advocatesFrontiers in Big Data10.3389/fdata.2024.14698197Online publication date: 20-Nov-2024
  • (2023)A Framework for Early Detection of Cyberbullying in Chinese-English Code-Mixed Social Media Text Using Natural Language Processing and Machine Learning2023 5th International Conference on Natural Language Processing (ICNLP)10.1109/ICNLP58431.2023.00061(298-302)Online publication date: Mar-2023
  • (2023)Natural language inference model for customer advocacy detection in online customer engagementMachine Language10.1007/s10994-023-06476-w113:4(2249-2275)Online publication date: 29-Nov-2023

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICEMC '22: Proceedings of the 2022 International Conference on E-business and Mobile Commerce
May 2022
173 pages
ISBN:9781450397162
DOI:10.1145/3543106
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 22 August 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. financial opinion mining
  2. natural language inference
  3. transformer-based pre-trained language models

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICEMC 2022

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)108
  • Downloads (Last 6 weeks)12
Reflects downloads up to 24 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Credibility-based knowledge graph embedding for identifying social brand advocatesFrontiers in Big Data10.3389/fdata.2024.14698197Online publication date: 20-Nov-2024
  • (2023)A Framework for Early Detection of Cyberbullying in Chinese-English Code-Mixed Social Media Text Using Natural Language Processing and Machine Learning2023 5th International Conference on Natural Language Processing (ICNLP)10.1109/ICNLP58431.2023.00061(298-302)Online publication date: Mar-2023
  • (2023)Natural language inference model for customer advocacy detection in online customer engagementMachine Language10.1007/s10994-023-06476-w113:4(2249-2275)Online publication date: 29-Nov-2023

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media