Abstract
Implicit discourse relation recognition is an important sub-task in discourse parsing, which needs to infer the relation based on proper discourse comprehension. Recent studies on cognitive learning strategies have suggested that using mental imagery strategy will foster text comprehension, which could effectively improve the capability of learners’ reading. Therefore, we propose a novel Mental Imagery-driven Neural Networks (MINN) to enhance representation for implicit discourse relation recognition. It employs the multi-granularity imagery vectors generated by the arguments to capture the deeper semantic information of discourse at different scales. Specifically, we 1) encode the different granularities of arguments (i.e., phrases, sentences.) and generate the corresponding imagery vectors as mentally imagining images of text content; 2) fuse the argument representations and imagery vectors as sequence representations; 3) further adopt self-attention to mine the important interactions between the sequence representations to infer the discourse relations. Extensive experimental results on the Penn Discourse TreeBank (PDTB) show that our model achieves competitive results against several state-of-the-art systems.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Bai, H., Zhao, H.: Deep enhanced representation for implicit discourse relation recognition. In: Proceedings of the 2018 COLING, pp. 571–583 (2018)
Chen, J., Zhang, Q., Liu, P., Qiu, X., Huang, X.: Implicit discourse relation detection via a deep architecture with gated relevance network. In: Proceedings of the 2016 ACL, pp. 1726–1735 (2016)
Dai, Z., Huang, R.: Improving implicit discourse relation classification by modeling inter-dependencies of discourse units in a paragraph. In: Proceedings of the 2018 NAACL, pp. 141–151 (2018)
Ji, Y., Eisenstein, J.: One vector is not enough: entity-augmented distributional semantics for discourse relations. Trans. ACL 3, 329–344 (2015)
Kosslyn, S.M.: Image and Brain: The Resolution of the Imagery Debate. MIT press, Cambridge (1996)
Lan, M., Wang, J., Wu, Y., Niu, Z.Y., Wang, H.: Multi-task attention-based neural networks for implicit discourse relationship representation and identification. In: Proceedings of the 2017 EMNLP, pp. 1299–1308 (2017)
Lei, W., Wang, X., Liu, M., Ilievski, I., He, X., Kan, M.Y.: Swim: a simple word interaction model for implicit discourse relation recognition. In: Proceedings of the 2017 IJCAI, pp. 4026–4032 (2017)
Leutner, D., den Elzen-Rump, V., Leopold, C.: Self-regulated learning from science texts. In: Prenzel, M. (ed.) Studies on the Educational Quality of Schools. The Final Report on the DFG Priority Programme, pp. 221–238 (2007)
Leutner, D., Leopold, C., Sumfleth, E.: Cognitive load and science text comprehension: effects of drawing and mentally imagining text content. Comput. Hum. Behav. 25(2), 284–289 (2009)
Lin, Z., Kan, M.Y., Ng, H.T.: Recognizing implicit discourse relations in the Penn discourse treebank. In: Proceedings of the 2009 EMNLP, pp. 343–351 (2009)
Liu, Y., Li, S.: Recognizing implicit discourse relations via repeated reading: neural networks with multi-level attention. In: Proceedings of the 2016 EMNLP, pp. 1224–1233 (2016)
Liu, Y., Li, S., Zhang, X., Sui, Z.: Implicit discourse relation classification via multi-task neural networks. In: Proceedings of the 2016 AAAI, pp. 2750–2756 (2016)
Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013)
Peters, M., et al.: Deep contextualized word representations. In: Proceedings of the 2018 NAACL, pp. 2227–2237 (2018)
Pitler, E., Louis, A., Nenkova, A.A.: Automatic sense prediction for implicit discourse relations in text. In: Proceedings of the AFNLP, pp. 683–691 (2009)
Prasad, R., et al.: The Penn discourse treebank 2.0. In: Proceedings of the 2008 LREC (2008)
Qin, L., Zhang, Z., Zhao, H.: Implicit discourse relation recognition with context-aware character-enhanced embeddings. In: Proceedings of the 2016 COLING, pp. 1914–1924 (2016)
Vaswani, A., et al.: Attention is all you need. In: Proceedings of the 2017 NIPS, pp. 5998–6008 (2017)
Xu, T., et al.: AttnGAN: fine-grained text to image generation with attentional generative adversarial networks. In: Proceedings of the 2018 CVPR, pp. 1316–1324 (2018)
Zhang, B., Su, J., Xiong, D., Lu, Y., Duan, H., Yao, J.: Shallow convolutional neural network for implicit discourse relation recognition. In: Proceedings of the 2015 EMNLP, pp. 2230–2235 (2015)
Zhang, H., et al.: StackGAN: text to photo-realistic image synthesis with stacked generative adversarial networks. In: Proceedings of the 2017 ICCV, pp. 5907–5915 (2017)
Acknowledgments
We thank the anonymous reviewers for their valuable feedback. Our work is supported by the National Natural Science Foundation of China (61976154), and the Tianjin Natural Science Foundation (18JCYBJC15500).
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Wang, J., He, R., Guo, F., Han, Y. (2020). Mental Imagery-Driven Neural Network to Enhance Representation for Implicit Discourse Relation Recognition. In: Farkaš, I., Masulli, P., Wermter, S. (eds) Artificial Neural Networks and Machine Learning – ICANN 2020. ICANN 2020. Lecture Notes in Computer Science(), vol 12397. Springer, Cham. https://doi.org/10.1007/978-3-030-61616-8_56
Download citation
DOI: https://doi.org/10.1007/978-3-030-61616-8_56
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-61615-1
Online ISBN: 978-3-030-61616-8
eBook Packages: Computer ScienceComputer Science (R0)