作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2022, Vol. 48 ›› Issue (1): 99-104. doi: 10.19678/j.issn.1000-3428.0059917

• 人工智能与模式识别 • 上一篇    下一篇

基于改进图节点的图神经网络多跳阅读理解研究

舒冲1, 欧阳智2, 杜逆索1,2, 何庆2, 魏琴2   

  1. 1. 贵州大学 计算机科学与技术学院, 贵阳 550025;
    2. 贵州大学 贵州省大数据产业发展应用研究院, 贵阳 550025
  • 收稿日期:2020-11-05 修回日期:2021-01-04 发布日期:2021-01-08
  • 作者简介:舒冲(1996-),男,硕士研究生,主研方向为自然语言处理;欧阳智、杜逆索、何庆、魏琴,副教授、博士。
  • 基金资助:
    国家重点研发计划(2018YFB1004300);贵州省科学技术厅重大科技计划(黔科合重大专项字[2018]3002);贵州大学培育项目(黔科合平台人才[2017]5788)。

Research on Multi-Hop Reading Comprehension Based on Graph Neural Network with Improved Graph Nodes

SHU Chong1, OUYANG Zhi2, DU Nisuo1,2, HE Qing2, WEI Qin2   

  1. 1. College of Computer Science and Technology, Guizhou University, Guiyang 550025, China;
    2. Guizhou Big Data Academy, Guizhou University, Guiyang 550025, China
  • Received:2020-11-05 Revised:2021-01-04 Published:2021-01-08

摘要: 多跳阅读理解需要基于问题并在多个支撑文档中寻找相关信息进行跳跃式推理来回答问题。针对当前多跳阅读理解模型中所存在的实体图内缺乏关键问题信息以及信息冗余问题,提出一种基于改进图节点的图神经网络多跳阅读理解模型。采用基于指代词的实体提取方法提取实体,将提取到的实体基于问题关联实体构建实体图。对实体图中的节点进行编码预处理,通过门机制的图卷积网络模拟得到推理信息,计算推理信息与问题信息的双向注意力并进行结果预测。在WikiHop数据集上的实验结果表明,该模型在测试集上取得了73.1%的预测准确率,相比基于图神经网络、循环神经网络和注意力机制的多跳阅读理解模型准确率更高、泛化性能更强。

关键词: 多跳阅读理解, 实体图, 问题关联实体, 图卷积网络, 双向注意力机制

Abstract: Multi-hop reading comprehension requires searching for question-associated information in supporting documents to perform leaping reasoning to answer the question.The entity graphs of existing multi-hop reading comprehension models lack key information relevant to the question, but contain redundant information.To address this problem, we propose a multi-hop reading comprehension model based on a graph neural network with improved graph nodes.We employ a demonstrative pronoun-based method to extract entities, and use the extracted entities to build an entity graph based on the question-associated entities.After the nodes in the entity graph are preprocessed by encoding, we use the gated Graph Convolutional Network(GCN) to obtain the inference information through simulation.The bidirectional attention between reasoning and question information is calculated, and on this basis the answer is predicted.The experimental results on the WikiHop dataset show that the model achieves a prediction accuracy of 73.1% on the test set.Compared with the multi-hop reading comprehension model based on GCN, Recurrent Neural Network(RNN) and attention mechanism, the proposed model displays higher accuracy and stronger generalization performance.

Key words: multi-hop reading comprehension, entity graph, question-associated entity, Graph Convolutional Network(GCN), bidirectional attention mechanism

中图分类号: