作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2022, Vol. 48 ›› Issue (10): 81-87,94. doi: 10.19678/j.issn.1000-3428.0063198

• 人工智能与模式识别 • 上一篇    下一篇

基于Mogrifier LSTM的序列标注关系抽取方法

方义秋1, 刘飞1, 葛君伟2   

  1. 1. 重庆邮电大学 计算机科学与技术学院, 重庆 400065;
    2. 重庆邮电大学 软件工程学院, 重庆 400065
  • 收稿日期:2021-11-10 修回日期:2021-12-10 发布日期:2022-10-09
  • 作者简介:方义秋(1963—),女,副教授,主研方向为云计算、大数据;刘飞(通信作者),硕士研究生;葛君伟,教授、博士。
  • 基金资助:
    国家自然科学基金面上项目“面向‘谣言-辟谣-促谣’博弈关系的社交网络谣言信息传播机制研究”(62072066)。

Method for Sequence Tagging Relationship Extraction Based on Mogrifier LSTM

FANG Yiqiu1, LIU Fei1, GE Junwei2   

  1. 1. College of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China;
    2. School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
  • Received:2021-11-10 Revised:2021-12-10 Published:2022-10-09

摘要: 对文本中的上下文信息进行充分利用能够提高关系抽取性能,但当前多数基于深度学习的关系抽取方法仅捕获文本序列中的浅层特征信息,在长序列文本中易丢失上下文信息。针对传统LSTM中输入和隐藏状态之间相互独立且缺少信息交互的问题,建立一种基于Mogrifier LSTM的序列标注关系抽取模型。将由词嵌入、字符嵌入和位置嵌入构成的嵌入层结果输入Mogrifier LSTM层,该层通过在传统LSTM计算之前交替地让当前输入与之前隐藏状态进行多轮运算,以增强上下文交互能力。使用自注意力机制提高模型对重要特征的关注度,同时采用基于关系的注意力机制获取特定关系下的句子表示,从而解决实体关系重叠问题。在关系分类模块,利用Bi-LSTM进行序列标注,将句子中的每个词汇映射为相应的标签。实验结果表明,在NYT数据集上该模型的F1值达到0.841,优于HRL、OrderRL等模型,在存在SEO重叠类型的关系抽取中,F1值仍能达到0.745,所提模型能够在一定程度上解决关系重叠问题同时有效提升关系抽取性能。

关键词: 关系抽取, Mogrifier LSTM模型, 上下文交互, 注意力机制, 关系重叠

Abstract: Making full use of the context information in the text can improve the relationship extraction performance.However, most current methods for deep-learning-based relationship extraction only capture the shallow feature information in the text sequence, and it is easy to lose the context information in the long sequence of text.A sequence tagging relationship extraction model based on Mogrifier Long Short-Term Memory(LSTM) is established to solve the independent input and hidden states and lack of information interaction in the traditional LSTM.The results of the embedding layer composed of word embedding, character embedding, and position embedding are input into the Mogrifier LSTM layer, which alternately performs multiple operations between the current input and the previous hidden states before the traditional LSTM calculation to improve the context interaction ability.The self-attention mechanism is used to improve the attention of the model to the critical features, and the relationship-based attention mechanism is used to obtain the sentence representation in the specific relationship to solve the problem of entity relationship overlap.In the relationship classification module, Bi-LSTM is used for sequence tagging to map each word in a sentence to a corresponding label.The experimental results show that the F1 value of the proposed model on the NYT dataset reaches 0.841, which is better than those of the HRL and OrderRL models.In relationship extraction with SEO overlap types, the F1 value reaches 0.745.The proposed model can reasonably solve the relationship overlap problem and effectively improve relationship extraction performance.

Key words: relationship extraction, Mogrifier LSTM model, context interaction, attention mechanism, relationship overlap

中图分类号: