Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3652628.3652731acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicaiceConference Proceedingsconference-collections
research-article

DBMAT: Research on Chinese Named Entity Recognition Using the Dilated Bidirectional Multi-layer Attentive Transformer Fusion Model

Published: 23 May 2024 Publication History

Abstract

In the realm of Chinese Named Entity Recognition tasks, conventional models have frequently fallen short in adequately addressing linguistic features and recognizing the essential role of context. To address this challenge, our research presents a unique Chinese Named Entity Recognition model, referred to as the LERT-DBMAT-CRF model. Initially, the LERT pre-trained language model incorporates language-informed pretraining strategies to enrich the semantic attributes in textual data. Subsequently, we apply the DBMAT module, unifying bidirectional Long Short-Term Memory networks with residual dilated convolutional networks, coordinated through a multi-head additive attention mechanism. This approach enhances feature extraction by employing the Exponential Linear Unit function, thus enhancing the model's capacity to capture temporal and spatial information relevant to semantic features. Lastly, a Conditional Random Field layer is introduced to exploit contextual information for label prediction. Experimental results demonstrate the exceptional model performance, achieving impressive F1 scores of 97.38% and 96.55% on the Resume dataset and the MSRA dataset, respectively, surpassing the performance of current mainstream models.

References

[1]
HARMAN D. Information retrieval: the early years [J]. Foundations and Trends® in Information Retrieval, 2019, 13(5): 425-577.
[2]
SONG Q, WU Y, LIN P, Mining summaries for knowledge graph search [J]. IEEE Transactions on Knowledge and Data Engineering, 2018, 30(10): 1887-1900.
[3]
Wu Y, Schuster M, Chen Z, 2016. Google's neural machine translation system: Bridging the gap between human and machine translation [J]. arXiv:1609.08144. Retrieved from https://arxiv.org/abs/1609.08144.
[4]
BIM, ZHANG Q, ZUO M, Bi-directional Long Short Term Memory Model with Semantic Positional Attention for the Question Answering System [J]. Transactions on Asian and Low-Resource Language Information Processing, 2021, 20(5): 1-13.
[5]
GAYEN V, SARKAR K. 2014. An HMM based named entity recognition system for Indian languages: the JU system at ICON 2013[J]. arXiv:1405.7397. Retrieved from https://arxiv.org/abs/1405.7397.
[6]
Schuldt C, Laptev I, Caputo B. Recognizing human actions: a local SVM approach [C]//Proceedings of the 17th International Conference on Pattern Recognition, 2004. I-CPR 2004. IEEE, 2004, 3: 32-36.
[7]
Sutton C, McCallum A. An introduction to conditional random fields [J]. Foundations and Trends® in Machine Learning, 2012, 4(4): 267-373.
[8]
Guillaume Lample, Miguel Ballesteros, Sandeep Subramanian, Neural architectures for named entity recognition [C]//North American Chapter of the Association f-or Computational Linguistics, 2016:387-396.
[9]
Chiu J P C, Nichols E. Named entity recognition with bi-directional LSTM-CNNs [J]. Transactions of the association for computational linguistics, 2016, 4: 357-370.
[10]
Strubell E, Verga P, Belanger D, 2017. Fast and accurate entity recognition with iterated dilated convolutions [J]. arXiv:1702.02098. Retrieved from https://arxiv.org/abs/1702.02098.
[11]
Devlin J, Chang M W, Lee K, 2018. Bert: Pretraining of deep bidirectional transformers for language understanding [J]. arXiv:1810.04805. Retrieved from https://arxiv.org/abs/1810.04805.
[12]
Chang Y, Kong L, Jia K, Chinese named entity recognition method based on BERT [C]//2021 IEEE International Conference on Data Science and Computer Application (ICDSCA). IEEE, 2021: 294-299.
[13]
Wang X Y, Li R, Duan F, An Efficient Chinese Named Entity Recognition Method Based on Gated-Dilated Convolution [J]. Journal of Chinese Information Processing, 2021, 35(1):72-80.
[14]
Tan Y J, Chen W, Yin Z. 2022. Chinese Named Entity Recognition for Gated-Dilated Convolution and Cascading Networks [J/OL]. Journal of Chinese Computer Systems. Retrieved from https://kns.cnki.net/kcms/detail/21.1106.tp.20220418.1445.032.html.
[15]
Cui Y, Che W, Wang S, 2022. Lert: A linguistically-motivated pre-trained language model [J]. arXiv:2211.05344. Retrieved from https://arxiv.org/abs/2211.05344.
[16]
Zhang Y, Yang J. 2018. Chinese NER using lattice LSTM [J]. arXiv:1805.02023. Retrieved from https://arxiv.org/abs/1805.02023.
[17]
Li X, Yan H, Qiu X, 2020. FLAT: Chinese NER using flat-lattice transformer [J]. arXiv:2004.11795. Retrieved from https://arxiv.org/abs/2004.11795.
[18]
Liu W, Fu X, Zhang Y, 2021. Lexicon enhanced Chinese sequence labeling using BERT adapter [J]. arXiv:2105.07148. Retrieved from https://arxiv.org/abs/2105.07148.
[19]
Zhu E, Li J. 2022. Boundary smoothing for named entity recognition [J]. arXiv:2204.12031. Retrieved from https://arxiv.org/abs/2204.12031.
[20]
Wu S, Song X, Feng Z, 2022. Nflat: Non-flat-lattice trans-former for chinese named entity recognition [J]. arXiv:2205.05832. Retrieved from https://arxiv.org/abs/2205.05832.
[21]
Xiong L, Zhou J, Zhu Q, 2023. A Confidence-based Parti-al Label Learning Model for Crowd-Annotated Named Entity Recognition [J]. arXiv:2305.12485. Retrieved from https://arxiv.org/abs/2305.12485.

Cited By

View all
  • (2024)Based on Gated Dynamic Encoding Optimization, the LGE-Transformer Method for Low-Resource Neural Machine TranslationIEEE Access10.1109/ACCESS.2024.348818612(162861-162869)Online publication date: 2024

Index Terms

  1. DBMAT: Research on Chinese Named Entity Recognition Using the Dilated Bidirectional Multi-layer Attentive Transformer Fusion Model

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICAICE '23: Proceedings of the 4th International Conference on Artificial Intelligence and Computer Engineering
    November 2023
    1263 pages
    ISBN:9798400708831
    DOI:10.1145/3652628
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 23 May 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICAICE 2023

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)5
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 23 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Based on Gated Dynamic Encoding Optimization, the LGE-Transformer Method for Low-Resource Neural Machine TranslationIEEE Access10.1109/ACCESS.2024.348818612(162861-162869)Online publication date: 2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media