计算机工程 ›› 2020, Vol. 46 ›› Issue (6): 60-64.doi: 10.19678/j.issn.1000-3428.0054540

• 人工智能与模式识别 • 上一篇    下一篇

基于双编码器结构的文本自动摘要研究

冯读娟, 杨璐, 严建峰   

  1. 苏州大学 计算机科学与技术学院, 江苏 苏州 215006
  • 收稿日期:2019-04-09 修回日期:2019-06-11 发布日期:2019-06-22
  • 作者简介:冯读娟(1994-),女,硕士研究生,主研方向为自然语言处理、自动摘要;杨璐、严建峰,副教授。
  • 基金项目:
    国家自然科学基金(61572339,61272449);江苏省科技支撑计划重点项目(BE2014005)。

Research on Automatic Text Summarization Based on Dual-Encoder Structure

FENG Dujuan, YANG Lu, YAN Jianfeng   

  1. School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
  • Received:2019-04-09 Revised:2019-06-11 Published:2019-06-22

摘要: 为了解决序列到序列模型中编码器不能充分编码源文本的问题,构建一种基于双编码器网络结构的CGAtten-GRU模型。2个编码器分别使用卷积神经网络和双向门控循环单元,源文本并行进入双编码器,结合2种编码网络结构的输出结果构建注意力机制,解码器端使用GRU网络融合Copy机制和集束搜索方法,以提高解码的准确度。在大规模中文短文本摘要数据集LCSTS上的实验结果表明,与RNN context模型相比,该模型的Rouge-1、Rouge-2和Rouge-L分别提高0.1、0.059和0.046。

关键词: 自然语言处理, 生成式摘要, 卷积神经网络, 门控循环单元, 注意力机制, 序列到序列模型, Copy机制

Abstract: This paper constructs a CGAtten-GRU model based on dual-encoder network structure to solve the problem that the encoder cannot fully encode the source text in the sequence-to-sequence(seq2seq) model.The two encoders use Convolutional Neural Network(CNN) and Bidirectional Gated Recurrent Unit(BiGRU) respectively,and the source text enters the two encoders in parallel.An attention mechanism is constructed by means of the outputs of two encoding networks.The decoder uses GRU network combining the Copy mechanism and the beam search method to improve the accuracy of decoding.Experimental results on large-scale Chinese short text summarization dataset LCSTS show that compared with the RNN context model,the proposed model improves Rouge-1 by 0.1,Rouge-2 by 0.059,and Rouge-L by 0.046.

Key words: Natural Language Processing(NLP), abstractive summarization, Convolutional Neural Network(CNN), Gated Recurrent Unit(GRU), attention mechanism, sequence-to-sequence(seq2seq) model, Copy mechanism

中图分类号: