Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3650400.3650478acmotherconferencesArticle/Chapter ViewAbstractPublication PageseitceConference Proceedingsconference-collections
research-article

LlmRe: A zero-shot entity relation extraction method based on the large language model

Published: 17 April 2024 Publication History

Abstract

Entity relation extraction aims to extract knowledge triples from unstructured or semi-structured text data and can be applied to various fields, including medicine, finance knowledge graph construction and intelligent question-answering. Traditional entity relation extraction requires a large amount of labeled data, consumes a lot of labor and time, and the trained model lacks generalization ability, which is difficult to migrate to other fields. Zero-shot entity relation extraction relieves the dependence on labeled data in traditional method. Based on unlabeled text data, zero-shot entity relation extraction has strong domain adaptability, which is a very challenging and practical task. Recent work on large language models shows that large models can effectively complete downstream tasks through natural language instructions and have good generalization ability. Inspired by this, we explore the use of large models for information extraction. Due to the randomness of large language model generation, we introduce in-context learning in entity relation extraction task to guide large language model to output data in a specified format to help obtain structured data. At the same time, we propose a three-stage extraction framework for decomposing entity relation extraction tasks, and each stage is conducted in the form of question and answer to reduce the complexity of extraction. We evaluated the knowledge triples extraction performance of the model on three self-built test datasets in different fields, and the experimental result showed that our proposed method achieved impressive performance in the zero-shot entity relation extraction task, surpassing the comparison model on multiple metrics, proving the effectiveness and domain adaptability of the proposed method.

References

[1]
Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., ... & Amodei, D. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901.
[2]
Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., ... & Lowe, R. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744.
[3]
Chinchor, N., & Marsh, E. 1998, July. Muc-7 information extraction task definition. In Proceeding of the seventh message understanding conference (MUC-7), Appendices (pp. 359-367).
[4]
Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. science, 313(5786), 504-507.
[5]
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosukhin, I. 2017. Attention is all you need. Advances in neural information processing systems, 30.
[6]
Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
[7]
Sun, Y., Wang, S., Li, Y., Feng, S., Chen, X., Zhang, H., ... & Wu, H. 2019. Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:1904.09223.
[8]
Zhong, Z., & Chen, D. 2020. A frustratingly easy approach for entity and relation extraction. arXiv preprint arXiv:2010.12812.
[9]
Wei, Z., Su, J., Wang, Y., Tian, Y., & Chang, Y. 2019. A novel cascade binary tagging framework for relational triple extraction. arXiv preprint arXiv:1909.03227.
[10]
Wang, Y., Yu, B., Zhang, Y., Liu, T., Zhu, H., & Sun, L. 2020. TPLinker: Single-stage joint extraction of entities and relations through token pair linking. arXiv preprint arXiv:2010.13415.
[11]
Zeng, X., Zeng, D., He, S., Liu, K., & Zhao, J. 2018, July. Extracting relational facts by an end-to-end neural model with copy mechanism. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (pp. 506-514).
[12]
Zeng, D., Liu, K., Chen, Y., & Zhao, J. 2015, September. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of the 2015 conference on empirical methods in natural language processing (pp. 1753-1762).
[13]
Lu, Y., Liu, Q., Dai, D., Xiao, X., Lin, H., Han, X., ... & Wu, H. 2022. Unified structure generation for universal information extraction. arXiv preprint arXiv: 2203.12277.
[14]
Wei, X., Cui, X., Cheng, N., Wang, X., Zhang, X., Huang, S., ... & Han, W. 2023. Zero-shot information extraction via chatting with chatgpt. arXiv preprint arXiv: 2302.10205.
[15]
Zeng, A., Liu, X., Du, Z., Wang, Z., Lai, H., Ding, M., ... & Tang, J. 2022. Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414.

Cited By

View all
  • (2024)SFDG-RE: Self-Feedback Description Generation Based on LLMs for Enhanced Zero-Shot Relation ExtractionAdvanced Data Mining and Applications10.1007/978-981-96-0814-0_25(383-398)Online publication date: 13-Dec-2024

Index Terms

  1. LlmRe: A zero-shot entity relation extraction method based on the large language model

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    EITCE '23: Proceedings of the 2023 7th International Conference on Electronic Information Technology and Computer Engineering
    October 2023
    1809 pages
    ISBN:9798400708305
    DOI:10.1145/3650400
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 17 April 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    EITCE 2023

    Acceptance Rates

    Overall Acceptance Rate 508 of 972 submissions, 52%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)153
    • Downloads (Last 6 weeks)41
    Reflects downloads up to 31 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)SFDG-RE: Self-Feedback Description Generation Based on LLMs for Enhanced Zero-Shot Relation ExtractionAdvanced Data Mining and Applications10.1007/978-981-96-0814-0_25(383-398)Online publication date: 13-Dec-2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media