Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3534678.3539443acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Public Access

Few-Shot Fine-Grained Entity Typing with Automatic Label Interpretation and Instance Generation

Published: 14 August 2022 Publication History

Abstract

We study the problem of few-shot Fine-grained Entity Typing (FET), where only a few annotated entity mentions with contexts are given for each entity type. Recently, prompt-based tuning has demonstrated superior performance to standard fine-tuning in few-shot scenarios by formulating the entity type classification task as a ''fill-in-the-blank'' problem. This allows effective utilization of the strong language modeling capability of Pre-trained Language Models (PLMs). Despite the success of current prompt-based tuning approaches, two major challenges remain: (1) the verbalizer in prompts is either manually designed or constructed from external knowledge bases, without considering the target corpus and label hierarchy information, and (2) current approaches mainly utilize the representation power of PLMs, but have not explored their generation power acquired through extensive general-domain pre-training. In this work, we propose a novel framework for few-shot FET consisting of two modules: (1) an entity type label interpretation module automatically learns to relate type labels to the vocabulary by jointly leveraging few-shot instances and the label hierarchy, and (2) a type-based contextualized instance generator produces new instances based on given instances to enlarge the training set for better generalization. On three benchmark datasets, our model outperforms existing methods by significant margins.

References

[1]
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, T. J. Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeff Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language Models are Few-Shot Learners. ArXiv, Vol. abs/2005.14165 (2020).
[2]
Shuang Chen, Jinpeng Wang, Feng Jiang, and Chin-Yew Lin. 2020 b. Improving Entity Linking by Modeling Latent Entity Type Information. In AAAI.
[3]
Tongfei Chen, Yunmo Chen, and Benjamin Van Durme. 2020 a. Hierarchical Entity Typing via Multi-level Learning to Rank. In ACL.
[4]
Xiang Chen, Ningyu Zhang, Lei Li, Xin Xie, Shumin Deng, Chuanqi Tan, Fei Huang, Luo Si, and Huajun Chen. 2021 b. LightNER: A Lightweight Generative Framework with Prompt-guided Attention for Low-resource NER. ArXiv, Vol. abs/2109.00720 (2021).
[5]
Xiang Chen, Ningyu Zhang, Xin Xie, Shumin Deng, Yunzhi Yao, Chuanqi Tan, Fei Huang, Luo Si, and Huajun Chen. 2021 c. KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction. ArXiv, Vol. abs/2104.07650 (2021).
[6]
Yi Chen, Haiyun Jiang, Lemao Liu, Shuming Shi, Chuang Fan, Min Yang, and Ruifeng Xu. 2021 a. An Empirical Study on Multiple Information Sources for Zero-Shot Fine-Grained Entity Typing. In EMNLP.
[7]
Kevin Clark, Minh-Thang Luong, Quoc V Le, and Christopher D Manning. 2020. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In ICLR.
[8]
Hongliang Dai, Donghong Du, Xin Li, and Yangqiu Song. 2019. Improving Fine-grained Entity Typing with Entity Linking. In EMNLP.
[9]
Hongliang Dai, Yangqiu Song, and Xin Li. 2020. Exploiting Semantic Relations for Fine-grained Entity Typing. In AKBC.
[10]
Hongliang Dai, Yangqiu Song, and Haixun Wang. 2021. Ultra-Fine Entity Typing with Weak Supervision from a Masked Language Model. In ACL/IJCNLP.
[11]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT.
[12]
Ning Ding, Yulin Chen, Xu Han, Guangwei Xu, Pengjun Xie, Haitao Zheng, Zhiyuan Liu, Juan-Zi Li, and Hong-Gee Kim. 2021 a. Prompt-Learning for Fine-Grained Entity Typing. ArXiv, Vol. abs/2108.10604 (2021).
[13]
Ning Ding, Guangwei Xu, Yulin Chen, Xiaobin Wang, Xu Han, Pengjun Xie, Haitao Zheng, and Zhiyuan Liu. 2021 b. Few-NERD: A Few-Shot Named Entity Recognition Dataset. In ACL/IJCNLP.
[14]
Xin Dong, Evgeniy Gabrilovich, Geremy Heitz, Wilko Horn, N. Lao, Kevin P. Murphy, Thomas Strohmann, Shaohua Sun, and Wei Zhang. 2014. Knowledge vault: a web-scale approach to probabilistic knowledge fusion. Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining (2014).
[15]
Anthony Fader, Luke Zettlemoyer, and Oren Etzioni. 2014. Open question answering over curated and extracted knowledge bases. Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining (2014).
[16]
Tianyu Gao, Adam Fisch, and Danqi Chen. 2021. Making Pre-trained Language Models Better Few-shot Learners. (2021).
[17]
Daniel Gillick, Nevena Lazic, Kuzman Ganchev, Jesse Kirchner, and David Huynh. 2014. Context-Dependent Fine-Grained Entity Type Tagging. ArXiv, Vol. abs/1412.1820 (2014).
[18]
Yuxian Gu, Xu Han, Zhiyuan Liu, and Minlie Huang. 2021. PPT: Pre-trained Prompt Tuning for Few-shot Learning. ArXiv, Vol. abs/2109.04332 (2021).
[19]
Xu Han, Weilin Zhao, Ning Ding, Zhiyuan Liu, and Maosong Sun. 2021. PTR: Prompt Tuning with Rules for Text Classification. ArXiv, Vol. abs/2105.11259 (2021).
[20]
Jeremy Howard and Sebastian Ruder. 2018. Universal Language Model Fine-tuning for Text Classification. In ACL.
[21]
Linmei Hu, Tianchi Yang, Chuan Shi, Houye Ji, and Xiaoli Li. 2019. Heterogeneous Graph Attention Networks for Semi-supervised Short Text Classification. In EMNLP.
[22]
Shengding Hu, Ning Ding, Huadong Wang, Zhiyuan Liu, Juan-Zi Li, and Maosong Sun. 2021. Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text Classification. ArXiv, Vol. abs/2108.02035 (2021).
[23]
Jiaxin Huang, Chunyuan Li, Krishan Subudhi, Damien Jose, Shobana Balakrishnan, Weizhu Chen, Baolin Peng, Jianfeng Gao, and Jiawei Han. 2021. Few-Shot Named Entity Recognition: A Comprehensive Study.
[24]
Zhengbao Jiang, Antonios Anastasopoulos, J. Araki, Haibo Ding, and Graham Neubig. 2020 a. X-FACTR: Multilingual Factual Knowledge Retrieval from Pretrained Language Models. In EMNLP.
[25]
Zhengbao Jiang, Frank F. Xu, J. Araki, and Graham Neubig. 2020 b. How Can We Know What Language Models Know? Transactions of the Association for Computational Linguistics, Vol. 8 (2020), 423--438.
[26]
Hailong Jin, Lei Hou, Juan-Zi Li, and Tiansi Dong. 2019. Fine-Grained Entity Typing via Hierarchical Multi Graph Convolutional Networks. In EMNLP.
[27]
Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In ICLR.
[28]
Dong-Ho Lee, Mahak Agarwal, Akshen Kadakia, Jay Pujara, and Xiang Ren. 2021. Good Examples Make A Faster Learner: Simple Demonstration-based Learning for Low-resource NER. ArXiv, Vol. abs/2110.08454 (2021).
[29]
Chen Liang, Yue Yu, Haoming Jiang, Siawpeng Er, Ruijia Wang, Tuo Zhao, and Chao Zhang. 2020. BOND: BERT-Assisted Open-Domain Named Entity Recognition with Distant Supervision. Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (2020).
[30]
Thomas Lin, Mausam, and Oren Etzioni. 2012. No Noun Phrase Left Behind: Detecting and Typing Unlinkable Entities. In EMNLP.
[31]
Ying Lin and Heng Ji. 2019. An Attentive Fine-Grained Entity Typing Model with Latent Type Representation. In EMNLP.
[32]
Xiao Ling and Daniel S. Weld. 2012. Fine-Grained Entity Recognition. In AAAI .
[33]
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv preprint arXiv:1907.11692 (2019).
[34]
Michal Lukasik, Srinadh Bhojanapalli, Aditya Krishna Menon, and Surinder Kumar. 2020. Does label smoothing mitigate label noise?
[35]
Yukun Ma, E. Cambria, and Sa Gao. 2016. Label Embedding for Zero-shot Fine-grained Named Entity Typing. In COLING.
[36]
Yu Meng, Yunyi Zhang, Jiaxin Huang, Xuan Wang, Yu Zhang, Heng Ji, and Jiawei Han. 2021. Distantly-Supervised Named Entity Recognition with Noise-Robust Learning and Language Model Augmented Self-Training. In EMNLP.
[37]
Yu Meng, Yunyi Zhang, Jiaxin Huang, Chenyan Xiong, Heng Ji, Chao Zhang, and Jiawei Han. 2020 a. Text Classification Using Label Names Only: A Language Model Self-Training Approach. In EMNLP.
[38]
Yu Meng, Yunyi Zhang, Jiaxin Huang, Yu Zhang, Chao Zhang, and Jiawei Han. 2020 b. Hierarchical Topic Mining via Joint Spherical Tree and Text Embedding. Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (2020).
[39]
Shikhar Murty, Pat Verga, Luke Vilnis, Irena Radovanovic, and Andrew McCallum. 2018. Hierarchical Losses and New Resources for Fine-grained Entity Typing and Linking. In ACL.
[40]
Rasha Obeidat, Xiaoli Z. Fern, Hamed Shahbazi, and Prasad Tadepalli. 2019. Description-Based Zero-shot Fine-Grained Entity Typing. In NAACL.
[41]
Fabio Petroni, Tim Rocktäschel, Patrick Lewis, Anton Bakhtin, Yuxiang Wu, Alexander H. Miller, and Sebastian Riedel. 2019. Language Models as Knowledge Bases?. In EMNLP .
[42]
Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners.
[43]
Xiang Ren, Wenqi He, Meng Qu, Lifu Huang, Heng Ji, and Jiawei Han. 2016a. AFET: Automatic Fine-Grained Entity Typing by Hierarchical Partial-Label Embedding. In EMNLP.
[44]
Xiang Ren, Wenqi He, Meng Qu, Clare R. Voss, Heng Ji, and Jiawei Han. 2016b. Label Noise Reduction in Entity Typing by Heterogeneous Partial-Label Embedding. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2016).
[45]
Timo Schick and Hinrich Schütze. 2021. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. In EACL .
[46]
Sonse Shimaoka, Pontus Stenetorp, Kentaro Inui, and Sebastian Riedel. 2017. Neural Architectures for Fine-grained Entity Type Classification. In EACL.
[47]
Taylor Shin, Yasaman Razeghi, Robert L Logan IV, Eric Wallace, and Sameer Singh. 2020. Eliciting Knowledge from Language Models Using Automatically Generated Prompts. ArXiv, Vol. abs/2010.15980 (2020).
[48]
Ralph Weischedel and Ada Brunstein. 2005. BBN Pronoun and Coreference and Entity Type Corpus. Linguistic Data Consortium (2005).
[49]
Ralph Weischedel, Martha Palmer, Mitchell Marcus, Eduard Hovy, Sameer Pradhan, Lance Ramshaw, and Nianwen Xue et al. 2013. OntoNotes Release 5.0. Linguistic Data Consortium (2013).
[50]
Ji Xin, Yankai Lin, Zhiyuan Liu, and Maosong Sun. 2018. Improving Neural Fine-Grained Entity Typing With Knowledge Attention. In AAAI.
[51]
Yadollah Yaghoobzadeh, Heike Adel, and Hinrich Schütze. 2017. Noise Mitigation for Neural Entity Typing and Relation Extraction. In EACL.
[52]
Mohamed Amir Yosef, Sandro Bauer, Johannes Hoffart, Marc Spaniol, and Gerhard Weikum. 2012. HYENA: Hierarchical Type Classification for Entity Names. In COLING.
[53]
Zheng Yuan and Doug Downey. 2018. OTyper: A Neural Architecture for Open Named Entity Typing. In AAAI.
[54]
Haoyu Zhang, Dingkun Long, Guangwei Xu, Muhua Zhu, Pengjun Xie, Fei Huang, and J. Wang. 2020 a. Learning with Noise: Improving Distantly-Supervised Fine-grained Entity Typing via Automatic Relabeling. In IJCAI.
[55]
T. Zhang, Congying Xia, Chun-Ta Lu, and Philip S. Yu. 2020 b. MZET: Memory Augmented Zero-Shot Fine-grained Named Entity Typing. In COLING.
[56]
Tony Zhao, Eric Wallace, Shi Feng, Dan Klein, and Sameer Singh. 2021. Calibrate Before Use: Improving Few-Shot Performance of Language Models. ArXiv, Vol. abs/2102.09690 (2021).
[57]
Ben Zhou, Daniel Khashabi, Chen-Tse Tsai, and Dan Roth. 2018. Zero-Shot Open Entity Typing as Type-Compatible Grounding. In EMNLP.

Cited By

View all
  • (2024)Dual Contrastive Learning for Cross-Domain Named Entity RecognitionACM Transactions on Information Systems10.1145/367887942:6(1-33)Online publication date: 18-Oct-2024
  • (2024)Few-Shot Name Entity Recognition on StackOverflow2024 9th International Conference on Intelligent Computing and Signal Processing (ICSP)10.1109/ICSP62122.2024.10743392(961-965)Online publication date: 19-Apr-2024
  • (2024)PRONTO: Prompt-Based Detection of Semantic Containment Patterns in MLMsThe Semantic Web – ISWC 202410.1007/978-3-031-77850-6_13(227-246)Online publication date: 11-Nov-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '22: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
August 2022
5033 pages
ISBN:9781450393850
DOI:10.1145/3534678
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 August 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. entity typing
  2. few-shot learning
  3. prompt-based learning

Qualifiers

  • Research-article

Funding Sources

  • AI Research Institutes program supported by NSF
  • Institute for Geospatial Understanding through an Integrative Discovery Environment (I-GUIDE) by NSF
  • Microsoft Research PhD Fellowship
  • US DARPA KAIROS Program
  • INCAS Program
  • Google PhD Fellowship
  • National Science Foundation

Conference

KDD '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Upcoming Conference

KDD '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)293
  • Downloads (Last 6 weeks)39
Reflects downloads up to 15 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Dual Contrastive Learning for Cross-Domain Named Entity RecognitionACM Transactions on Information Systems10.1145/367887942:6(1-33)Online publication date: 18-Oct-2024
  • (2024)Few-Shot Name Entity Recognition on StackOverflow2024 9th International Conference on Intelligent Computing and Signal Processing (ICSP)10.1109/ICSP62122.2024.10743392(961-965)Online publication date: 19-Apr-2024
  • (2024)PRONTO: Prompt-Based Detection of Semantic Containment Patterns in MLMsThe Semantic Web – ISWC 202410.1007/978-3-031-77850-6_13(227-246)Online publication date: 11-Nov-2024
  • (2023)A Hierarchical Prototype Contrastive Learning based Few-Shot Entity Typing ApproachProceedings of the 2023 International Conference on Information Education and Artificial Intelligence10.1145/3660043.3660198(876-881)Online publication date: 22-Dec-2023
  • (2023)Pretrained Language Representations for Text Understanding: A Weakly-Supervised PerspectiveProceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3580305.3599569(5817-5818)Online publication date: 6-Aug-2023
  • (2023)Decoupled Hyperbolic Graph Attention Network for Cross-domain Named Entity RecognitionProceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3539618.3591662(591-600)Online publication date: 19-Jul-2023
  • (2022)Adapting Pretrained Representations for Text MiningProceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3534678.3542607(4806-4807)Online publication date: 14-Aug-2022

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media