Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3640912.3640993acmotherconferencesArticle/Chapter ViewAbstractPublication PagescnmlConference Proceedingsconference-collections
research-article

LAL-JER: Label-Aware Learning for Adaptive Joint Entity and Relation Extraction with LLM data augmentation

Published: 22 February 2024 Publication History

Abstract

Joint entity and relation extraction has achieved great improvements in Natural Language Processing (NLP) and has been widely applied, such as constructing knowledge graph, query understanding and question answering. Existing methods usually spend long time on fitting the models on certain datasets with given label type, which greatly lacks the ability of generalization. The model cannot make prediction on label types that have not seen in the training set. To address this issue, we propose to use prompt to incorporate the semantic meaning of the label type description. Furthermore, we use large language model to perform data augmentation to improve the robustness of our model during training. Extensive experiments and ablation study on two joint entity and relation extraction validates the effectiveness of our work on that: 1. Our methods achieved states of art performance on joint entity and relation extraction benchmark based on pretrained language model bert. 2. Our methods can help the model make predictions on label type unseen before given prompts.

References

[1]
D. Zelenko, C. Aone, and A. Richardella. "Kernel Methods for Relation Extraction." Journal of Machine Learning Research, 3: 1083-1106, 2003.
[2]
Y. S. Chan and D. Roth. "Exploiting Syntactico-Semantic Structures for Relation Extraction." In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, 551-560. Portland, Oregon, USA: Association for Computational Linguistics, 2011.
[3]
Q. Li and H. Ji. "Incremental Joint Extraction of Entity Mentions and Relations." In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 402-412. Baltimore, Maryland: Association for Computational Linguistics, 2014.
[4]
Z. Wei, J. Su, Y. Wang, "A novel cascade binary tagging framework for relational triple extraction." arXiv preprint arXiv:1909.03227, 2019.
[5]
Y. Wang, B. Yu, Y. Zhang, "TPLinker: Single-stage joint extraction of entities and relations through token pair linking." arXiv preprint arXiv:2010.13415, 2020.
[6]
P. Gupta, H. Schütze, and B. Andrassy. "Table filling multitask recurrent neural network for joint entity and relation extraction." In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, 2537-2547, 2016.
[7]
Y. M. Shang, H. Huang, and X. Mao. "Onerel: Joint entity and relation extraction with one module in one step." In Proceedings of the AAAI Conference on Artificial Intelligence, 36(10): 11285-11293, 2022.
[8]
J. Wang and W. Lu. "Two are better than one: Joint entity and relation extraction with table-sequence encoders." arXiv preprint arXiv:2010.03851, 2020.
[9]
W. Tang, B. Xu, Y. Zhao, "UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction." arXiv preprint arXiv:2211.09039, 2022.
[10]
J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova. "BERT: Pre-training of deep bidirectional transformers for language understanding." In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171-4186, 2019.
[11]
Y. Liu, M. Ott, N. Goyal, "Roberta: A robustly optimized bert pretraining approach." arXiv preprint arXiv:1907.11692, 2019.
[12]
P. He, X. Liu, J. Gao, "Deberta: Decoding-enhanced bert with disentangled attention." arXiv preprint arXiv:2006.03654, 2020.
[13]
Y. Lu, Q. Liu, D. Dai, "Unified structure generation for universal information extraction." arXiv preprint arXiv:2203.12277, 2022.
[14]
J. Yu, B. Bohnet, and M. Poesio. "Named entity recognition as dependency parsing." arXiv preprint arXiv:2005.07150, 2020.
[15]
J. Su. "GPLinker: Entity and Relation Extraction based on GlobalPointer." Kexue.fm, Jan 2022. [Online]. Available: https://kexue.fm/archives/8888.
[16]
H. Touvron, T. Lavril, G. Izacard, "Llama: Open and efficient foundation language models." arXiv preprint arXiv:2302.13971, 2023.
[17]
L. Zintgraf, K. Shiarli, V. Kurin, K. Hofmann, and S. Whiteson. "Fast Context Adaptation via Meta-Learning." In K. Chaudhuri and R. Salakhutdinov (Eds.), * Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, 7693-7702, 2019.
[18]
S. Min, X. Lyu, A. Holtzman, "Rethinking the role of demonstrations: What makes in-context learning work?" arXiv preprint arXiv:2202.12837, 2022.
[19]
T. Nguyen and E. Wong. "In-context Example Selection with Influences." arXiv preprint arXiv:2302.11042, 2023.
[20]
C. Olsson, N. Elhage, N. Nanda, "In-context learning and induction heads." arXiv preprint arXiv:2209.11895, 2022.
[21]
X. Li and X. Qiu. "Finding supporting examples for in-context learning." arXiv preprint arXiv:2302.13539, 2023.
[22]
J. Gao, C. Yu, W. Wang, "Mask-then-fill: A flexible and effective data augmentation framework for event extraction." arXiv preprint arXiv:2301.02427, 2023.
[23]
S. Riedel, L. Yao, and A. McCallum. "Modeling relations and their mentions without labeled text." In ECML, 2010, pp. 148-163.
[24]
C. Gardent, A. Shimorina, S. Narayan, and L. Perez-Beltrachini. "Creating training corpora for NLG microplanners." In ACL, 2017, pp. 179-188.
[25]
Z. Zhong and D. Chen. "A frustratingly easy approach for entity and relation extraction." arXiv preprint arXiv:2010.12812, 2020.
[26]
J. Liu, S. Chen, B. Wang, J. Zhang, N. Li, and T. Xu. "Attention as relation: Learning supervised multi-head self-attention for relation extraction." In IJCAI, 2020, pp. 3787-3793.
[27]
B. Guo, S. Han, X. Han, "Label confusion learning to enhance text classification models." In Proceedings of the AAAI conference on artificial intelligence, 35(14): 12929-12936, 2021.
[28]
Y. Shou, T. Meng, W. Ai, "Conversational emotion recognition studies based on graph convolutional neural networks and a dependent syntactic analysis." Neurocomputing, 501: 629-639, 2022.
[29]
T. Meng, Y. Shou, W. Ai, J. Du, H. Liu, and K. Li. "A Multi-Message Passing Framework Based on Heterogeneous Graphs in Conversational Emotion Recognition." Available at SSRN 4353605, 2021.

Index Terms

  1. LAL-JER: Label-Aware Learning for Adaptive Joint Entity and Relation Extraction with LLM data augmentation
    Index terms have been assigned to the content through auto-classification.

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    CNML '23: Proceedings of the 2023 International Conference on Communication Network and Machine Learning
    October 2023
    446 pages
    ISBN:9798400716683
    DOI:10.1145/3640912
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 22 February 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    CNML 2023

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 68
      Total Downloads
    • Downloads (Last 12 months)68
    • Downloads (Last 6 weeks)10
    Reflects downloads up to 17 Nov 2024

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media