Abstract
Question answering (QA) is a challenging task in natural language processing. Recently, with the remarkable success of deep learning on many natural language processing tasks, including semantic and syntactic analysis, machine translation, relation extraction, etc., more and more efforts have also been devoted to the task of question answering. This chapter briefly introduces the recent advances in deep learning methods on two typical and popular question answering tasks. (1) Deep learning in question answering over knowledge base (KBQA) which mainly employs deep neural networks to understand the meaning of the questions and try to translate them into structured queries, or directly translate them into distributional semantic representations compared with candidate answers in the knowledge base. (2) Deep learning in machine comprehension (MC) which manages to construct an end-to-end paradigm based on novel neural networks for directly computing the deep semantic matching among question, answers and the given passage.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
A CVT node is usually not a real-world entity, but often refers to an event, e.g., a marriage event, or a presidency event, which can represent an entry of data with multiple fields.
- 2.
More details can be found in https://nlp.stanford.edu/software/sempre/.
- 3.
Obtained through https://www.microsoft.com/en-us/download/details.aspx?id=52763.
- 4.
Obtained through http://beforethecode.com/projects/omdb/download.aspx.
- 5.
Obtained through http://grouplens.org/datasets/movielens/.
- 6.
- 7.
- 8.
References
Berant, J., Chou, A., Frostig, R., & Liang, P. (2013). Semantic parsing on freebase from question-answer pairs. In EMNLP.
Berant, J., & Liang, P. (2014). Semantic parsing via paraphrasing. In ACL.
Bordes, A., Chopra, S., & Weston, J. (2014a). Question answering with subgraph embeddings. In EMNLP.
Bordes, A., Usunier, N., Chopra, S., & Weston, J. (2015). Large-scale simple question answering with memory networks. In arXiv.
Bordes, A., Weston, J., & Usunier, N. (2014b). Open question answering with weakly supervised embedding models. In ECML.
Cai, Q., & Yates, A. (2013). Large-scale semantic parsing via schema matching and lexicon extension. In ACL.
Chen, D., Bolton, J., & Manning, C. D. (2016). A thorough examination of the CNN/Daily Mail reading comprehension task. In Association for Computational Linguistics (ACL).
Cui, Y., Chen, Z., Wei, S., Wang, S., Liu, T., & Hu, G. (2017). Attention-over-attention neural networks for reading comprehension. In ACL.
Dhingra, B., Liu, H., Yang, Z., Cohen, W. W., & Salakhutdinov, R. (2016). Gated-attention readers for text comprehension. arXiv preprint arXiv:1606.01549.
Dong, L., Wei, F., Zhou, M., & Xu, K. (2015). Question answering over freebase with multi-column convolutional neural networks. In ACL-IJCNLP.
Etzioni, O. (2011). Search needs a shake-up. Nature, 476(7358), 25–26.
Hao, Y., Zhang, Y., Liu, K., He, S., Liu, Z., Wu, H., & Zhao, J. (2017). An end-to-end model for question answering over knowledge base with cross-attention combining global knowledge. In Association for Computational Linguistics (ACL).
Hermann, K. M., Kocisky, T., Grefenstette, E., Espeholt, L., Kay, W., Suleyman, M., & Blunsom, P. (2015). Teaching machines to read and comprehend. In Advances in Neural Information Processing Systems (pp. 1693–1701).
Hill, F., Bordes, A., Chopra, S., & Weston, J. (2015). The goldilocks principle: Reading children’s books with explicit memory representations. arXiv preprint arXiv:1511.02301.
Jansen, P., Surdeanu, M., & Clark, P. (2014). Discourse complements lexical semantics for non-factoid answer reranking. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Vol. 1: Long Papers, pp. 977–986). Association for Computational Linguistics.
Kobayashi, S., Tian, R., Okazaki, N., & Inui, K. (2016). Dynamic entity representation with max-pooling improves machine reading. In Proceedings of the 2016 Conference of the NAACL.
Kun, X., Sheng, Z., Yansong, F., & Dongyan, Z. (2014). Answering natural language questions via phrasal semantic parsing. In Proceedings of the 2014 Conference on Natural Language Processing and Chinese Computing (NLPCC).
Kwiatkowski, T., Choi, E., Artzi, Y., & Zettlemoyer, L. S. (2013). Scaling semantic parsers with on-the-fly ontology matching. In EMNLP.
Liang, C., Berant, J., Le, Q., Forbus, K. D., & Lao, N. (2017). Neural Symbolic Machines: Learning Semantic Parsers on Freebase with Weak Supervision. In Proceedings of the Association for Computational Linguistics (ACL 2017). Canada: Association for Computational Linguistics.
Liu, Y., Wei, F., Li, S., Ji, H., Zhou, M., & Wang, H. (2015). A dependency-based neural network for relation classification. In ACL.
Miller, A., Fisch, A., Dodge, J., Karimi, A.-H., Bordes, A., & Weston, J. (2016). Key-value memory networks for directly reading documents. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (pp. 1400–1409). Austin, TX: Association for Computational Linguistics.
Narasimhan, K., & Barzilay, R. (2015). Machine comprehension with discourse relations. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Vol. 1: Long Papers, pp. 1253–1262). Association for Computational Linguistics.
Rajpurkar, P., Zhang, J., Lopyrev, K., & Liang, P. (2016). Squad: 100, 000+ questions for machine comprehension of text. CoRR, abs/1606.05250.
Reddy, S., Lapata, M., & Steedman, M. (2014). Large-scale semantic parsing without question-answer pairs. Transactions of the Association of Computational Linguistics (pp. 377–392).
Richardson, M., Burges, J. C., & Renshaw, E. (2013). Mctest: A challenge dataset for the open-domain machine comprehension of text. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing (pp. 193–203). Association for Computational Linguistics.
Sachan, M., Dubey, K., Xing, E., & Richardson, M. (2015). Learning answer-entailing structures for machine comprehension. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Vol. 1: Long Papers, pp. 239–249). Association for Computational Linguistics.
Seo, M. J., Kembhavi, A., Farhadi, A., & Hajishirzi, H. (2016). Bidirectional attention flow for machine comprehension. CoRR, abs/1611.01603.
Shen, Y., Huang, P.-S., Gao, J., & Chen, W. (2017). Reasonet: Learning to stop reading in machine comprehension. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’17 (pp. 1047–1055). New York, USA: ACM.
Smith, E., Greco, N., Bosnjak, M., & Vlachos, A. (2015). A strong lexical matching method for the machine comprehension test. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (pp. 1693–1698). Association for Computational Linguistics.
Sordoni, A., Bachman, P., Trischler, A., & Bengio, Y. (2016). Iterative alternating neural attention for machine reading. arXiv preprint arXiv:1606.02245.
Steedman, M. (2000). The Syntactic Process. Cambridge, MA: The MIT Press.
Sugawara, S., Yokono, H., & Aizawa, A. (2017). Prerequisite skills for reading comprehension: Multi-perspective analysis of mctest datasets and systems.
Sukhbaatar, S., Weston, J., Fergus, R., et al. (2015). End-to-end memory networks. In Advances in Neural Information Processing Systems (pp. 2440–2448).
Taylor, W. L. (1953). cloze procedure: a new tool for measuring readability. Journalism Bulletin, 30(4), 415–433.
Trischler, A., Ye, Z., Yuan, X., & Suleman, K. (2016). Natural language comprehension with the epireader. arXiv preprint arXiv:1606.02270.
Wang, S., & Jiang, J. (2016). Machine comprehension using match-lstm and answer pointer. CoRR, abs/1608.07905.
Wang, W., Yang, N., Wei, F., Chang, B., & Zhou, M. (2017). Gated self-matching networks for reading comprehension and question answering. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Vol. 1: Long Papers, pp. 189–198). Association for Computational Linguistics.
Wang, Z., Mi, H., Hamza, W., & Florian, R. (2016). Multi-perspective context matching for machine comprehension. CoRR, abs/1612.04211.
Weston, J., Bordes, A., Chopra, S., Rush, A. M., van Merriënboer, B., Joulin, A., & Mikolov, T. (2015a). Towards ai-complete question answering: A set of prerequisite toy tasks. arXiv preprint arXiv:1502.05698.
Weston, J., Chopra, S., & Bordes, A. (2015b). Memory networks. In ICLR.
Xiong, C., Zhong, V., & Socher, R. (2016). Dynamic coattention networks for question answering. CoRR, abs/1611.01604.
Xu, K., Feng, Y., Huang, S., & Zhao, D. (2015). Semantic relation classification via convolutional neural networks with simple negative sampling. In EMNLP.
Xu, K., Reddy, S., Feng, Y., Huang, S., & Zhao, D. (2016). Question Answering on Freebase via Relation Extraction and Textual Evidence. In Proceedings of the Association for Computational Linguistics (ACL 2016). Berlin, Germany: Association for Computational Linguistics.
Yang, Y., & Chang, M.-W. (2015). S-mart: Novel tree-based structured learning algorithms applied to tweet entity linking. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Vol. 1: Long Papers, pp. 504–513). Beijing, China: Association for Computational Linguistics.
Yang, Z., Dhingra, B., Yuan, Y., Hu, J., Cohen, W. W., & Salakhutdinov, R. (2016). Words or characters? Fine-grained gating for reading comprehension. CoRR, abs/1611.01724.
Yih, W.-t., Chang, M.-W., He, X., & Gao, J. (2015). Semantic parsing via staged query graph generation: Question answering with knowledge base. In ACL-IJCNLP.
Yih, W.-t., He, X., & Meek, C. (2014). Semantic parsing for single-relation question answering. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Vol. 2: Short Papers, pp. 643–648). Baltimore, MD: Association for Computational Linguistics.
Yih, W.-t., Richardson, M., Meek, C., Chang, M.-W., & Suh, J. (2016). The value of semantic parse labeling for knowledge base question answering. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Vol. 2: Short Papers, pp. 201–206). Berlin, Germany: Association for Computational Linguistics.
Yu, Y., Zhang, W., Hasan, K. S., Yu, M., Xiang, B., & Zhou, B. (2016). End-to-end reading comprehension with dynamic answer chunk ranking. CoRR, abs/1610.09996.
Zeng, D., Liu, K., Lai, S., Zhou, G., & Zhao, J. (2014). Relation classification via convolutional deep neural network. In Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers (pp. 2335–2344). Dublin, Ireland: Dublin City University and Association for Computational Linguistics.
Zhang, S., Feng, Y., Huang, S., Xu, K., Han, Z., & Zhao, D. (2015). Semantic interpretation of superlative expressions via structured knowledge bases. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Vol. 2: Short Papers, pp. 225–230). Beijing, China: Association for Computational Linguistics.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Liu, K., Feng, Y. (2018). Deep Learning in Question Answering. In: Deng, L., Liu, Y. (eds) Deep Learning in Natural Language Processing. Springer, Singapore. https://doi.org/10.1007/978-981-10-5209-5_7
Download citation
DOI: https://doi.org/10.1007/978-981-10-5209-5_7
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-5208-8
Online ISBN: 978-981-10-5209-5
eBook Packages: Computer ScienceComputer Science (R0)