Abstract
Relation classification aims to recognize semantic relation between two given entities mentioned in the given text. Existing models have performed well on the inverse relation classification with large-scale datasets, but their performance drops significantly for few-shot learning. In this paper, we propose a Phrase-level Attention Network, function words adaptively enhanced attention framework (FAEA+), to attend class-related function words by the designed hybrid attention for few-shot inverse relation classification in Knowledge Graph. Then, an instance-aware prototype network is present to adaptively capture relation information associated with query instances and eliminate intra-class redundancy due to function words introduced. We theoretically prove that the introduction of function words will increase intra-class differences, and the designed instance-aware prototype network is competent for reducing redundancy. Experimental results show that FAEA+ significantly improved over strong baselines on two few-shot relation classification datasets. Moreover, our model has a distinct advantage in solving inverse relations, which outperforms state-of-the-art results by 16.82% under a 1-shot setting in FewRel1.0.
Similar content being viewed by others
Data Availability
The FewRel 1.0 and FewRel 2.0 datesets we used are available in https://thunlp.github.io/fewrel.html.
References
Dong, H., Wang, W., Coenen, F., Huang, K.: Knowledge base enrichment by relation learning from social tagging data. Inf. Sci. 526, 203–220 (2020). https://doi.org/10.1016/j.ins.2020.04.002
Lee, Y.: Improving end-to-end task-oriented dialog system with a simple auxiliary task. In: Findings of the Association for Computational Linguistics: EMNLP, pp. 1296–1303 (2021). https://doi.org/10.18653/v1/2021.findings-emnlp.112
Bhattacharya, I., Getoor, L.: Collective entity resolution in relational data. ACM Transactions on Knowledge Discovery from Data 1(1), 5 (2007)
Sridhar, D., Fakhraei, S., Getoor, L.: A probabilistic approach for collective similarity-based drug-drug interaction prediction. Bioinform. 32(20), 3175–3182 (2016)
Han, X., Zhu, H., Yu, P., Wang, Z., Yao, Y., Liu, Z., Sun, M.: Fewrel: A large-scale supervised few-shot relation classification dataset with state-of-the-art evaluation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 4803–4809 (2018). https://doi.org/10.18653/v1/d18-1514
Liu, Y., Hu, J., Wan, X., Chang, T.: A simple yet effective relation information guided approach for few-shot relation extraction. In: Findings of the Association for Computational Linguistics, pp. 757–763 (2022). https://doi.org/10.18653/v1/2022.findings-acl.62
Gao, T., Han, X., Liu, Z., Sun, M.: Hybrid attention-based prototypical networks for noisy few-shot relation classification. In: The Thirty-Third AAAI Conference on Artificial Intelligence, pp. 6407–6414 (2019). https://doi.org/10.1609/aaai.v33i01.33016407
Snell, J., Swersky, K., Zemel, R.S.: Prototypical networks for few-shot learning. In: Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems, pp. 4077–4087 (2017)
Gao, T., Han, X., Zhu, H., Liu, Z., Li, P., Sun, M., Zhou, J.: Fewrel 2.0: Towards more challenging few-shot relation classification. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pp. 6249–6254 (2019). https://doi.org/10.18653/v1/D19-1649
Sun, S., Sun, Q., Zhou, K., Lv, T.: Hierarchical attention prototypical networks for few-shot text classification. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pp. 476–485 (2019). https://doi.org/10.18653/v1/D19-1045
Yang, S., Zhang, Y., Niu, G., Zhao, Q., Pu, S.: Entity concept-enhanced few-shot relation extraction. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, pp. 987–991 (2021). https://doi.org/10.18653/v1/2021.acl-short.124
Bao, Y., Wu, M., Chang, S., Barzilay, R.: Few-shot text classification with distributional signatures. In: 8th International Conference on Learning Representations (2020)
Dou, C., Wu, S., Zhang, X., Feng, Z., Wang, K.: Function-words adaptively enhanced attention networks for few-shot inverse relation classification. In: Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, pp. 2937–2943 (2022). https://doi.org/10.24963/ijcai.2022/407
Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 4171–4186 (2019). https://doi.org/10.18653/v1/n19-1423
Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning, pp. 1126–1135 (2017)
Bo, D., Wang, X., Shi, C., Shen, H.: Beyond low-frequency information in graph convolutional networks. In: Thirty-Fifth AAAI Conference on Artificial Intelligence, pp. 3950–3957 (2021)
Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: 7th International Conference on Learning Representations (2019)
Rasooli, M.S., Tetreault, J.R.: Yara parser: A fast and accurate dependency parser. Computer Science 194(1), 93–106 (2015)
Peng, H., Gao, T., Han, X., Lin, Y., Li, P., Liu, Z., Sun, M., Zhou, J.: Learning from context or names? an empirical study on neural relation extraction. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pp. 3661–3672 (2020). https://doi.org/10.18653/v1/2020.emnlp-main.298
Ye, Z., Ling, Z.: Multi-level matching and aggregation network for few-shot relation classification. In: Proceedings of the 57th Conference of the Association for Computational Linguistics, pp. 2872–2881 (2019). https://doi.org/10.18653/v1/p19-1277
Satorras, V.G., Estrach, J.B.: Few-shot learning with graph neural networks. In: 6th International Conference on Learning Representations (2018)
Qu, M., Gao, T., Xhonneux, L.A.C., Tang, J.: Few-shot relation extraction via bayesian meta-learning on relation graphs. In: Proceedings of the 37th International Conference on Machine Learning, pp. 7867–7876 (2020)
Wen, W., Liu, Y., Ouyang, C., Lin, Q., Chung, T.L.: Enhanced prototypical network for few-shot relation extraction. Inf. Process. Manag. 58(4), 102596 (2021). https://doi.org/10.1016/j.ipm.2021.102596
Wang, Y., Verspoor, K., Baldwin, T.: Learning from unlabelled data for clinical semantic textual similarity. In: Proceedings of the 3rd Clinical Natural Language Processing Workshop, ClinicalNLP@EMNLP 2020, pp. 227–233 (2020). https://doi.org/10.18653/v1/2020.clinicalnlp-1.25
Han, J., Cheng, B., Lu, W.: Exploring task difficulty for few-shot relation extraction. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 2605–2616 (2021). https://doi.org/10.18653/v1/2021.emnlp-main.204
Liu, Y., Hu, J., Wan, X., Chang, T.: Learn from relation information: Towards prototype representation rectification for few-shot relation extraction. In: Findings of the Association for Computational Linguistics: NAACL, pp. 1822–1831 (2022). https://doi.org/10.18653/v1/2022.findings-naacl.139
Soares, L.B., FitzGerald, N., Ling, J., Kwiatkowski, T.: Matching the blanks: Distributional similarity for relation learning. In: Proceedings of the 57th Conference of the Association for Computational Linguistics, pp. 2895–2905 (2019). https://doi.org/10.18653/v1/p19-1279
Xie, Y., Xu, H., Li, J., Yang, C., Gao, K.: Heterogeneous graph neural networks for noisy few-shot relation classification. Knowl. Based Syst. 194, 105548 (2020). https://doi.org/10.1016/j.knosys.2020.105548
Li, Y., Ma, Z., Gao, L., Wu, Y., Xie, F., Ren, X.: Enhance prototypical networks with hybrid attention and confusing loss function for few-shot relation classification. Neurocomputing 493, 362–372 (2022). https://doi.org/10.1016/j.neucom.2022.04.067
Gregory, K., Richard, Z., Ruslan, S.: Siamese neural networks for one-shot image recognition. In: Proceedings of the 3rd International Conference on Learning Representation (2015)
Li, Z., Ouyang, F., Zhou, C., He, Y., Shen, L.: Few-shot relation classification research based on prototypical network and causal intervention. IEEE Access 10, 36995–37002 (2022). https://doi.org/10.1109/ACCESS.2022.3164688
Yang, K., Zheng, N., Dai, X., He, L., Huang, S., Chen, J.: Enhance prototypical network with text descriptions for few-shot relation classification. In: The 29th ACM International Conference on Information and Knowledge Management, pp. 2273–2276 (2020). https://doi.org/10.1145/3340531.3412153
Zhang, J., Zhu, J., Yang, Y., Shi, W., Zhang, C., Wang, H.: Knowledge-enhanced domain adaptation in few-shot relation classification. In: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 2183–2191 (2021). https://doi.org/10.1145/3447548.3467438
Tang, Y., Li, Z., Cao, C., Fang, F., Cao, Y., Liu, Y., Fu, J.: Knowledge-based diverse feature transformation for few-shot relation classification. In: Knowledge Science, Engineering and Management - 14th International Conference. Lecture Notes in Computer Science, vol. 12815, pp. 101–114 (2021). https://doi.org/10.1007/978-3-030-82136-4_9
Ravi, S., Larochelle, H.: Optimization as a model for few-shot learning. In: 5th International Conference on Learning Representations (2017)
Dong, B., Yao, Y., Xie, R., Gao, T., Han, X., Liu, Z., Lin, F., Lin, L., Sun, M.: Meta-information guided meta-learning for few-shot relation classification. In: Proceedings of the 28th International Conference on Computational Linguistics, COLING 2020, pp. 1594–1605 (2020). https://doi.org/10.18653/v1/2020.coling-main.140
Scarselli, F., Gori, M., Tsoi, A.C., Hagenbuchner, M., Monfardini, G.: The graph neural network model. IEEE Trans. Neural Networks 20(1), 61–80 (2009). https://doi.org/10.1109/TNN.2008.2005605
Yang, L., Li, L., Zhang, Z., Zhou, X., Zhou, E., Liu, Y.: DPGN: distribution propagation graph network for few-shot learning. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13387–13396 (2020). https://doi.org/10.1109/CVPR42600.2020.01340
Guo, Z., Zhang, Y., Lu, W.: Attention guided graph convolutional networks for relation extraction. In: Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, pp. 241–251 (2019). https://doi.org/10.18653/v1/p19-1024
Sun, K., Zhang, R., Mao, Y., Mensah, S., Liu, X.: Relation extraction with convolutional network over learnable syntax-transport graph. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, pp. 8928–8935 (2020)
Tian, Y., Chen, G., Song, Y., Wan, X.: Dependency-driven relation extraction with attentive graph convolutional networks. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, pp. 4458–4471 (2021). https://doi.org/10.18653/v1/2021.acl-long.344
Wang, Y., Zhang, H.: Introducing graph neural networks for few-shot relation prediction in knowledge graph completion task. In: Knowledge Science, Engineering and Management - 14th International Conference. Lecture Notes in Computer Science, vol. 12815, pp. 294–306 (2021). https://doi.org/10.1007/978-3-030-82136-4_24
Li, Z., Sun, Y., Zhu, J., Tang, S., Zhang, C., Ma, H.: Improve relation extraction with dual attention-guided graph convolutional networks. Neural Comput. Appl. 33(6), 1773–1784 (2021). https://doi.org/10.1007/s00521-020-05087-z
Zhao, H., Xie, J., Wang, H.: Graph convolutional network based on multi-head pooling for short text classification. IEEE Access 10, 11947–11956 (2022). https://doi.org/10.1109/ACCESS.2022.3146303
Yang, S., Cai, B., Cai, T., Song, X., Jiang, J., Li, B., Li, J.: Robust cross-network node classification via constrained graph mutual information. Knowl. Based Syst. 257, 109852 (2022). https://doi.org/10.1016/j.knosys.2022.109852
Fang, U., Li, J., Akhtar, N., Li, M., Jia, Y.: Gomic: Multi-view image clustering via self-supervised contrastive heterogeneous graph co-learning. World Wide Web, 1–17 (2022)
Acknowledgements
This work was supported by National Natural Science Foundation of China (NSFC) (61972455).
Author information
Authors and Affiliations
Contributions
Shaojuan Wu: Conceptualization, Methodology, Wrote the abstract, introduction and theoretical Analysis. Chunliu Dou: Data curation, Methodology, Validation and Wrote original draft preparation. Dazhuang Wang: Software, Validation, Wrote the experiment and related work. Jitong Li: Software, Validation and prepared some figures. Xiaowang Zhang: Supervision, Project administration. Zhiyong Feng: Supervision, Funding acquisition. Kewen Wang: Analysis logic, Reviewed the manuscript. Sofonias Yitagesu: Reviewed the manuscript.
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This article belongs to the Topical Collection: Special Issue on Knowledge-Graph-Enabled Methods and Applications for the Future Web Guest Editors: Xin Wang, Jeff Pan, Qingpeng Zhang, Yuan-Fang Li.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wu, S., Dou, C., Wang, D. et al. Phrase-level attention network for few-shot inverse relation classification in knowledge graph. World Wide Web 26, 3001–3026 (2023). https://doi.org/10.1007/s11280-023-01142-6
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11280-023-01142-6