作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2020, Vol. 46 ›› Issue (2): 59-64,71. doi: 10.19678/j.issn.1000-3428.0053545

• 人工智能与模式识别 • 上一篇    下一篇

基于多层注意力机制的回指消解算法

刘雨江1,2, 付立军1,2, 刘俊明1,2, 吕鹏飞3   

  1. 1. 中国科学院大学 计算机科学与技术学院, 北京 100049;
    2. 中国科学院沈阳计算技术研究所 研究生部, 沈阳 110168;
    3. 中国地质图书馆信息技术研究中心, 北京 100083
  • 收稿日期:2019-01-02 修回日期:2019-03-04 发布日期:2019-03-14
  • 作者简介:刘雨江(1994-),男,硕士研究生,主研方向为自然语言处理;付立军、刘俊明,教授;吕鹏飞,高级工程师。
  • 基金资助:
    国土资源部大数据科研专项(201511079-3)。

Anaphora Resolution Algorithm Based on Multilayer Attention Mechanism

LIU Yujiang1,2, FU Lijun1,2, LIU Junming1,2, Lü Pengfei3   

  1. 1. School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing 100049, China;
    2. Graduate Faculty, Shenyang Institute of Computing Technology, University of Chinese Academy of Sciences, Shenyang 110168, China;
    3. Research Center of Information Technology, National Geological Library of China, Beijing 100083, China
  • Received:2019-01-02 Revised:2019-03-04 Published:2019-03-14

摘要: 在信息抽取过程中,无法被判别的回指易造成信息抽取不完整的情况,这种指代关系可通过分析当前语境下的指代部分、被指代部分、周围的信息及原文内容生成的唯一判别信息进行判断。为此,构建一个多层注意力机制模型,在不同层次上对上述信息进行基于注意力机制的概率计算,利用最终结果判别回指关系是否成立。在指代部分与被指代部分向量化后,通过2个注意力层上的4次概率计算,使每一个训练结果在判别之前都具有唯一性。在OntoNotes 5.0数据集上的实验结果表明,该模型F值在显性指代和零指代均存在的条件下为70.1%,在存在零指代的条件下为60.7%,高于尹庆宇等人提出的模型。

关键词: 指代关系, 注意力机制, 显性指代, 零指代, 多层注意力机制模型

Abstract: In the information extraction process,the nondeterministic anaphora can cause incomplete information extraction.By analyzing the only discriminate information generated by the anaphoric part,the referenced part,the surrounding information,the referenced surrounding information and the original content in the current context,the anaphora relations are judged and a multilayer attention mechanism model is constructed.The probability calculation based on attention mechanism is performed on these five parts at different levels,and the final results are used to determine whether the anaphora relations can be proved or not.With the vectorization of the anaphoric part and the referenced part,the four probability calculations on two attention layers make every training result unique before judgment.Experimental results on OntoNotes 5.0 dataset show that the F value of the proposed model is 70.1% when both overt anaphora and zero anaphora are presented.When only zero anaphora are presented,the F value is 60.7%,which is higher than the model proposed by YIN Qingyu et al.

Key words: anaphora relations, attention mechanism, overt anaphora, zero anaphora, multilayer attention mechanism model

中图分类号: