The knowledge graph-based question answering system aims to find the connection between the internal knowledge behind the problem and the known knowledge base triplet. This paper proposes a framework for knowledge graph question answering (ZDNN-KGQA), which divides intelligent question answering based on knowledge graphs into four steps: named entity recognition, entity disambiguation, attribute classification, and answer selection. In the named entity recognition step, this paper proposes a BERT-BiLSTM-CRF model that combines an attention mechanism to identify entities in question sentences. In the entity disambiguation step, this paper utilizes the BERT-Softmax model to extract the accurate entities from the candidate entity set. In the attribute classification step, it is converted into a problem of text similarity. This paper proposes the use of BERT-Softmax and Word2vec models to extract attributes intext problems, effectively improving entity recognition performance and question semantic parsing performance. The average F1 value of this system on the test dataset provided by the NLPCC-ICCPOL 2016 KBQA task is 0.8974, which is close to the optimal performance level.
|