Abstract
With the advent of natural language enabled applications, there has been a growing appetite for conversational question answering systems. This demand is being largely satisfied with the help of such powerful language models as Open AI’s GPT models, Google’s BERT, and BigScience’s BLOOM. However, the astounding amount of training data and computing resources required to create such models is a huge challenge. Furthermore, for such systems, catering to multiple application domains typically requires the acquisition of even more training data. We discuss an alternative approach to the problem of open-domain conversational question answering by utilizing knowledge graphs to capture relevant information from a body of text in any domain. We achieve this by allowing the relations of the knowledge graphs to be drawn directly from the body of text being processed, rather than from a fixed ontology. By connecting this process with SPARQL queries generated from natural language questions, we demonstrate the foundations of an open-domain question answering system that requires no training and can switch domains flexibly and seamlessly.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Angeli, G., Johnson, M., Manning, C.D.: Leveraging linguistic structure for open domain information extraction. In: Annual Meeting of the Association for Computational Linguistics (2015)
Brown, T.B., et al.: Language models are few-shot learners. arXiv abs/2005.14165 (2020)
Brożek, A.: The Structure of Natural Language Questions, pp. 129–169. Brill, Leiden, The Netherlands (2011)
Dai, Z., Li, L., Xu, W.: CFO: conditional focused neural question answering with large-scale knowledge bases. ArXiv abs/1606.01994 (2016)
Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein, J., Doran, C., Solorio, T. (eds.) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2–7, 2019, Volume 1 (Long and Short Papers), pp. 4171–4186 (2019)
Gangemi, A., Presutti, V., Recupero, D.R., Nuzzolese, A.G., Draicchio, F., Mongiovì, M.: Semantic web machine reading with FRED. Semant. Web 8(6), 873–893 (2017)
Hirschman, L., Gaizauskas, R.J.: Natural language question answering: the view from here. Nat. Lang. Eng. 7(4), 275–300 (2001)
Jamil, H., Oduro-Afriyie, J.: Semantic understanding of natural language stories for near human question answering. In: FQAS (2019)
Kim, K., Hur, Y., Kim, G., Lim, H.: GREG: a global level relation extraction with knowledge graph embedding. Appl. Sci. 10, 1181 (2020)
Lin, S.C., Hilton, J., Evans, O.: TruthfulQA: measuring how models mimic human falsehoods. In: Annual Meeting of the Association for Computational Linguistics (2021)
Miller, G.A.: WordNet: a lexical database for English. Commun. ACM 38, 39–41 (1992)
Pradel, C., Haemmerlé, O., Hernandez, N.: Natural language query interpretation into SPARQL using patterns. In: Fourth International Workshop on Consuming Linked Data - COLD (2013)
Scao, T.L., et al.: BLOOM: A 176b-parameter open-access multilingual language model. CoRR abs/2211.05100 (2022)
Sima, A.C., et al.: Bio-SODA: enabling natural language question answering over knowledge graphs without training data. In: Zhu, Q., Zhu, X., Tu, Y., Xu, Z., Kumar, A. (eds.) SSDBM 2021: 33rd International Conference on Scientific and Statistical Database Management, Tampa, FL, USA, July 6–7, 2021, pp. 61–72. ACM (2021)
Sobieszek, A., Price, T.: Playing games with Ais: The limits of GPT-3 and similar large language models. Minds Mach. 32, 341–364 (2022)
Vakulenko, S., Longpre, S., Tu, Z., Anantha, R.: Question rewriting for conversational question answering. In: Proceedings of the 14th ACM International Conference on Web Search and Data Mining (2021)
Vaswani, A., et al.: Attention is all you need. In: NIPS (2017)
You, C., Chen, N., Zou, Y.: Contextualized attention-based knowledge transfer for spoken conversational question answering. In: Interspeech (2021)
Zeng, D., Liu, K., Chen, Y., Zhao, J.: Distant supervision for relation extraction via piecewise convolutional neural networks. In: EMNLP (2015)
Zhao, F., Hou, J., Li, Y., Bai, L.: Relation prediction for answering natural language questions over knowledge graphs. In: International Joint Conference on Neural Networks, IJCNN 2021, Shenzhen, China, July 18–22, 2021, pp. 1–8. IEEE (2021)
Zou, L., Huang, R., Wang, H., Yu, J.X., He, W., Zhao, D.: Natural language question answering over RDF: a graph data driven approach. In: Proceedings of the 2014 ACM SIGMOD International Conference on Management of Data (2014)
Acknowledgement
This publication was partially made possible by an Institutional Development Award (IDeA) from the National Institute of General Medical Sciences of the National Institutes of Health under Grant #P20GM103408.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 Springer Nature Switzerland AG
About this paper
Cite this paper
Oduro-Afriyie, J., Jamil, H. (2023). Knowledge Graph Enabled Open-Domain Conversational Question Answering. In: Larsen, H.L., Martin-Bautista, M.J., Ruiz, M.D., Andreasen, T., Bordogna, G., De Tré, G. (eds) Flexible Query Answering Systems. FQAS 2023. Lecture Notes in Computer Science(), vol 14113. Springer, Cham. https://doi.org/10.1007/978-3-031-42935-4_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-42935-4_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-42934-7
Online ISBN: 978-3-031-42935-4
eBook Packages: Computer ScienceComputer Science (R0)