Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3468891.3468918acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmltConference Proceedingsconference-collections
research-article

Improving Thai Named Entity Recognition Performance Using BERT Transformer on Deep Networks

Published: 06 September 2021 Publication History

Abstract

Emerging of deep learning and transformer model helps advance many NLP tasks. For Name Entity Recognition (NER), many studies have applied deep transformer architect with Google BERT and ELMO to English language and achieved significant performance improvement compared to traditional embeddings. However, currently there is very little research on applying BERT transformer to Thai NER task, so in this paper we explore two different approaches to apply BERT transform to improve Thai NER which are fine-tuning BERT and using BERT as embedding. We found that Bi-LSTM-CRF network with BERT embedding model achieved a significant performance improvement with averaged F1 Score of 94%, without character embedding feature. This approach yields better performance than using only fine-tuning BERT for Thai NER and better than state of the art on Thai NER from Bi-LSTM with Thai2fit word embedding and character embedding.

References

[1]
J. Li, A. Sun, J. Han and C. Li, "A Survey on Deep Learning for Named Entity Recognition," IEEE Transactions on Knowledge and Data Engineering, pp. 1-1, 2020.
[2]
J. Devlin, M.-W. Chang, K. Lee and K. Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding," Association for Computational Linguistics, p. 4171–4186, 6 2019.
[3]
Y. Liu, F. Meng, J. Zhang, J. Xu, Y. Chen and J. Zhou, GCDT: A Global Context Enhanced Deep Transition Architecture for Sequence Labeling, Florence, Italy: Association for Computational Linguistics, 2019, p. 2431–2441.
[4]
Y. Luo, F. Xiao and H. Zhao, Hierarchical Contextualized Representation for Named Entity Recognition, 2019.
[5]
N. Tirasaroj, Thai Named Entity Regognition: The Application of Conditional Random Fields Models, 2010.
[6]
C. Udomcharoenchaikit, P. Vateekul and P. Boonkwan, "Thai Named-Entity Recognition Using Variational Long Short-Term Memory with Conditional Random Field," Advances in Intelligent Informatics, Smart Technology and Natural Language Processing. iSAI-NLP, vol. 807, 2017.
[7]
S. Thattinaphanich and S. Prom-on, "Thai Named Entity Recognition Using Bi-LSTM-CRF with Word and Character Representation," 2019 4th International Conference on Information Technology (InCIT), pp. 149-154, 2019.
[8]
L. Lowphansirikul, C. Polpanumas, N. Jantrakulchai and S. Nutanong, WangchanBERTa: Pretraining transformer-based Thai Language Models, arXiv:2101.09635, 2021.
[9]
Y. Wu, M. Schuster, Z. Chen, Q. V. Le, M. Norouzi, W. Macherey, M. Krikun, Y. Cao, Q. Gao, K. Macherey, J. Klingner, A. Shah, M. Johnson, X. Liu, Ł. Kaiser, S. Gouws, Y. Kato, T. Kudo, H. Kazawa, K. Stevens, G. Kurian, N. Patil, W. Wang, C. Young, J. Smith, J. Riesa, A. Rudnick, O. Vinyals, G. Corrado, M. Hughes and J. Dean, Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation, 2016.
[10]
ThAIKeras, "ThAIKeras/bert," [Online]. Available: https://github.com/ThAIKeras/bert.
[11]
C. Polpanumas, "cstorm125/thai2fit," [Online]. Available: https://github.com/cstorm125/thai2fit.
[12]
T. Kudo and J. Richardson, "SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing," in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Brussels, 2018.
[13]
W. Phatthiyaphaibun., "Thai Named Entity Recognitions for PyThaiNLP," [Online]. Available: https://github.com/wannaphong/thai-ner.
[14]
M. E. Peters, S. Ruder and N. A. Smith, "To Tune or Not to Tune? Adapting Pretrained Representations to Diverse Tasks," in Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019), Florence, 2019.
[15]
Q. Guo, X. Qiu, P. Liu, Y. Shao, X. Xue and Z. Zhang, "Star-Transformer," in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, 2019.
[16]
L. Martin, B. Muller, P. J. Ortiz Suárez, Y. Dupont, L. Romary, É. de la Clergerie, D. Seddah and B. Sagot, "CamemBERT: a Tasty French Language Model," in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 2020.

Cited By

View all
  • (2024)Named Entity Recognition for Thai Historical Data2024 21st International Joint Conference on Computer Science and Software Engineering (JCSSE)10.1109/JCSSE61278.2024.10613644(528-533)Online publication date: 19-Jun-2024
  • (2022)Ethically Responsible Machine Learning in FintechIEEE Access10.1109/ACCESS.2022.320288910(97531-97554)Online publication date: 2022

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICMLT '21: Proceedings of the 2021 6th International Conference on Machine Learning Technologies
April 2021
183 pages
ISBN:9781450389402
DOI:10.1145/3468891
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 06 September 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. BERT
  2. Bi-LSTM
  3. Conditional Random Filed
  4. Name Entity Recognition
  5. Transformer

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICMLT 2021

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)27
  • Downloads (Last 6 weeks)0
Reflects downloads up to 18 Aug 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Named Entity Recognition for Thai Historical Data2024 21st International Joint Conference on Computer Science and Software Engineering (JCSSE)10.1109/JCSSE61278.2024.10613644(528-533)Online publication date: 19-Jun-2024
  • (2022)Ethically Responsible Machine Learning in FintechIEEE Access10.1109/ACCESS.2022.320288910(97531-97554)Online publication date: 2022

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media