Abstract
Classical Chinese poetry, as the cultural heritage of human beings, is very popular in Chinese community all over the world. Nearly every person in these regions can recite several poems to artistically express his or her emotion. However, due to the huge difference between the archaic and modern Chinese language, now, it is hard for people to understand these poems. Considering the Neural machine translation (NMT) has made great progress since the self-attention technology called Transformer was adopted, we followed the paradigm of NMT to translate the classical Chinese poems into modern Chinese texts in this study. First, we collect many ancient poems and their explanations as a parallel corpus. Then, a NMT system is trained with this corpus. Finally, we compared this system with Generative Pre-Training (GPT) and analysis their results in detail.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Zhang, X., Lapata, M.: Chinese poetry generation with recurrent neural networks. In: EMNLP, pp. 670–680 (2014)
Yang, X., Lin X., Suo, S., Li, M.: Generating thematic Chinese poetry using conditional variational autoencoders with hybrid decoders. In: IJCAI, pp. 4539–4545 (2018)
Liao, Y., Wang Y., Liu, Q., Jiang, X.: GPT-based generation for classical Chinese poetry. In: arxiv (2019)
Sutskever, I., Vinyals, O., Le, Q.: Sequence to sequence learning with neural networks. In: NIPS, pp. 2238–2244 (2014)
Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)
Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language Models are Unsupervised Multitask Learners. Technical report (2019)
Papineni, K., Roukos S., Ward, T., Zhu, W.: BLEU: a method for automatic evaluation of machine translation. In: ACL (2002)
Yan, R.: I, Poet: automatic poetry composition through recurrent neural networks with iterative polishing schema. In: IJCAI, pp. 2238–2244 (2016)
Yi, X., Li R., Sun, M.: Chinese poetry generation with a salient-clue mechanism. In: Proceedings of the SIGNLL Conference on Computational Natural Language Learning, pp. 241–250 (2018)
Bahdanau, D., Cho K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: ICLR (2015)
Li, J., et al.: Generating classical Chinese poems via conditional variational autoencoder and adversarial training. In: EMNLP, pp. 3890–3900 (2018)
Xu, W., Ritter A., Dolan, W., Grishman, R., Cherry, C.: Paraphrasing for style. In: COLING (2012)
Jang, S., Min J., Kwon, M. : Writing Style Conversion using Neural Machine Translation. Technical report (2017)
Tian, H.: Learning to interpret and translate classical Chinese poetry. In: Theory and Practice in Language Studies. vol. 3, no. 9, pp. 1546–1551. Springer, Heidelberg (2013). https://doi.org/10.4304/tpls.3.9
Devlin, J., Chang M., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: NAACL-HLT, pp. 4171–4186 (2019)
Yi, X., Sun M., Li, R., Li, W.: Automatic poetry generation with mutual reinforcement learning. In: ACL (2018)
He, D., et al.: Dual learning for machine translation. In: NIPS (2016)
Hazra, R., Dutta P., Gupta, S., Qaathir, M., Dukkipati, A.: Active2 learning: actively reducing redundancies in active learning methods for sequence tagging and machine translation. In: NAACL (2021)
Acknowledgments
This study is supported by the fundamental research of Leshan Normal University (grant no. 801 205190117) and Natural Science Foundation of China (no. 61373056). We will thank the anonymous reviewers for their valuable comments.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Jin, P., Wang, H., Ma, L., Wang, B., Zhu, S. (2022). Translating Classical Chinese Poetry into Modern Chinese with Transformer. In: Dong, M., Gu, Y., Hong, JF. (eds) Chinese Lexical Semantics. CLSW 2021. Lecture Notes in Computer Science(), vol 13249. Springer, Cham. https://doi.org/10.1007/978-3-031-06703-7_37
Download citation
DOI: https://doi.org/10.1007/978-3-031-06703-7_37
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-06702-0
Online ISBN: 978-3-031-06703-7
eBook Packages: Computer ScienceComputer Science (R0)