Abstract
In response to the challenge of existing relation triplet extraction models struggling to adapt to new relation categories in zero-shot scenarios, we propose a method that combines generated synthetic training data with the retrieval of relevant documents through a rank-based filtering approach for data augmentation. This approach alleviates the problem of low-quality synthetic training data and reduces noise that may affect the accuracy of triplet extraction in certain relation categories. Experimental results on two public datasets demonstrate that our model exhibits stable and impressive performance compared to the baseline models in terms of precision, recall, and F1 score, resulting in improved effectiveness for zero-shot relation triplet extraction.
This work is supported by the National Natural Science Foundation of China (61972003), R &D Program of Beijing Municipal Education Commission (KM202210009002), the Beijing Urban Governance Research Base of North China University of Technology (2023CSZL16), and the North China University of Technology Startup Fund.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Nayak, T., Majumder, N., Goyal, P., et al.: Deep neural approaches to relation triplets extraction: a comprehensive survey. Cogn. Comput. 13(5), 1215–1232 (2021)
Chen, C.-Y., Li, C.-T.: ZS-BERT: towards zero-shot relation extraction with attribute representation learning. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 3470–3479 (2021)
Chia, Y.K., Bing, L., Poria, S., Si, L.: RelationPrompt: leveraging prompts to generate synthetic data for zero-shot relation triplet extraction. In: Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, pp. 45–57 (2022)
Lewis, P., Perez, E., Piktus, A., et al.: Retrieval-augmented generation for knowledge-intensive NLP tasks. In: Advances in Neural Information Processing Systems (NeurIPS 2020), vol. 33, pp. 9459–9474 (2020)
Karpukhin, V., Oguz, B., Min, S., et al.: Dense passage retrieval for open-domain question answering. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 6769–6781 (2020)
Petroni, F., Rocktäschel, T., Lewis, P., et al.: Language models as knowledge bases? In: Proceedings of EMNLP-IJCNLP (2019)
Keskar, N.S., McCann, B., Varshney, L.R., et al.: CTRL: a conditional transformer language model for controllable generation, arXiv preprint arXiv:1909.05858 (2019)
Yang, P., Li, L., Luo, F., et al.: Enhancing topic-to-essay generation with external commonsense knowledge. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 2002–2012 (2019)
Cabot, P.-L.H., Navigli, R.: REBEL: relation extraction by end-to-end language generation, In: Findings of the Association for Computational Linguistics: EMNLP 2021, Punta Cana, Dominican Republic, pp. 2370–2381 (2021)
Kim, B., Iso, H., Bhutani, N., Hruschka, E., Nakashole, N.: Zero-shot triplet extraction by template infilling. arXiv preprint arXiv:2212.10708 (2022)
Johnson, J., Douze, M., Jégou, H.: Billion-scale similarity search with GPUS. IEEE Trans. Big Data 7(3), 535–547 (2019)
Malkov, Y.A., Yashunin, D.A.: Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs. IEEE Trans. Pattern Anal. Mach. Intell. 42(4), 824–836 (2016)
Han, X., et al.: FewRel: a large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, pp. 4803–4809 (2018)
Wang, J., Lu, W.: Two are better than one: joint entity and relation extraction with table-sequence encoders. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pp. 1706–1721. Association for Computational Linguistics (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, Q. et al. (2024). Zero-Shot Relation Triplet Extraction via Retrieval-Augmented Synthetic Data Generation. In: Luo, B., Cheng, L., Wu, ZG., Li, H., Li, C. (eds) Neural Information Processing. ICONIP 2023. Communications in Computer and Information Science, vol 1969. Springer, Singapore. https://doi.org/10.1007/978-981-99-8184-7_28
Download citation
DOI: https://doi.org/10.1007/978-981-99-8184-7_28
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8183-0
Online ISBN: 978-981-99-8184-7
eBook Packages: Computer ScienceComputer Science (R0)