作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程

• 人工智能及识别技术 • 上一篇    下一篇

基于字符级循环网络的查询意图识别模型

孟奎 1,刘梦赤 1,胡婕 2   

  1. (1.武汉大学 计算机学院,武汉 430072; 2.湖北大学 计算机与信息工程学院,武汉 430062)
  • 收稿日期:2016-03-15 出版日期:2017-03-15 发布日期:2017-03-15
  • 作者简介:孟奎(1990—),男,硕士,主研方向为数据挖掘、数据库;刘梦赤,教授、博士生导师;胡婕,副教授。
  • 基金资助:
    国家自然科学基金(61202100)。

Query Intention Recognition Model Based on Character Level Cyclic Network

MENG Kui  1,LIU Mengchi  1,HU Jie  2   

  1. (1.Computer School, Wuhan University,Wuhan 430072,China; 2.School of Computer Science and Information Engineering,Hubei University,Wuhan 430062,China)
  • Received:2016-03-15 Online:2017-03-15 Published:2017-03-15

摘要: 以特征模板为主的查询意图识别方法存在人工抽取特征繁琐,并且难以捕捉文本语义信息的问题。为此,基于字符级循环网络,提出一种新的查询意图识别模型。为能有效提取句子深层次语义特征,减少长距离信息依赖的限制,使用长短时记忆网络(LSTM)作为神经网络线性变换层,同时增加一层反向LSTM抽取字符的将来信息特征。使用原始汉字直接作为模型的输入,避免分词结果不准确带来的错误传导问题,利用字符的分布向量表示方法,提高句子语义特征的获取。实验结果表明,该方法整体准确率达到90.7%,相比特征模板方法有所提升,能提高用户查询意图的分类性能。

关键词: 查询意图, 字符级, 循环神经网络, 记忆网络, 词向量

Abstract: Intention recognition methods,which are mainly based on feature template,have complicated hand-crafted feature extraction process and are difficult to capture semantic information of texts.Aiming at this problem,this paper proposes a new query intention identification model based on character level recurrent network.In order to effectively extract deep semantic features of a sentence and decrease long distance information dependent constraints,this paper uses Long Short -Term Memory Neural Network(LSTM) as a linear transformation of neural network layer,and uses a reverse LSTM layer to extract future information character.To avoid error propagation problem caused by inaccurate word segmentation results,it uses Chinese characters as inputs of the model,and uses distributed representation of characters to improve extractions of semantic features of sentences.Experimental results show that the method has an accuracy of 90.7%,which is higher than the characteristics template method,and it can improve the classification performance of user query intention.

Key words: query intention, character level, Recurrent Neural Network(RNN), memory network, word vector

中图分类号: