作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2023, Vol. 49 ›› Issue (7): 125-134. doi: 10.19678/j.issn.1000-3428.0065117

• 人工智能与模式识别 • 上一篇    下一篇

基于多元信息嵌入与协同神经网络的力学实体识别算法

张家熔1, 苑津莎1, 许珈宁2, 罗志宏3   

  1. 1. 华北电力大学 电子与通信工程系, 河北 保定 071000
    2. 国网阜新供电公司 电力调度控制中心, 辽宁 阜新 123000
    3. 华北电力大学 自动化系, 河北 保定 071000
  • 收稿日期:2022-06-30 出版日期:2023-07-15 发布日期:2022-09-22
  • 作者简介:

    张家熔(1992—),男,博士研究生,主研方向为深度学习、智能解答算法、智能信息处理

    苑津莎,教授、博士生导师

    许珈宁,硕士

    罗志宏,博士研究生

  • 基金资助:
    中央高校基本科研业务费专项资金(2020JG006)

Mechanics Entities Recognition Algorithm Based on Multi-Meta Information Embedding and Collaborative Neural Network

Jiarong ZHANG1, Jinsha YUAN1, Jianing XU2, Zhihong LUO3   

  1. 1. Department of Electronic and Communication Engineering, North China Electric Power University, Baoding 071000, Hebei, China
    2. Power Dispatching Control Center, State Grid Fuxin Power Supply Company, Fuxin 123000, Liaoning, China
    3. Department of Automation, North China Electric Power University, Baoding 071000, Hebei, China
  • Received:2022-06-30 Online:2023-07-15 Published:2022-09-22

摘要:

自动抽取力学问题中的关键实体是力学问题自动解答的重要手段。然而,与开放域相比,力学问题具有大量的专业词汇和较长的实体,其识别难度高,准确率低。针对该问题,基于图注意力网络(GAT)和Transformer编码器提出一种用于力学问题关键实体抽取的实体识别算法。针对汉语的特点设计一个包括字符信息、词汇信息和部首信息的多元信息嵌入用于增强中文句子表示。提出结构图和语境图两个图模型对中文句子进行建模,并设计一种协同架构,该架构使用两个独立的GAT整合多元信息并学习句子的上下文信息。为平衡词汇信息与部首信息对中文字符的影响,提出一种协同Transformer架构,该架构由字符-词汇Transformer与字符-部首Transformer构成,并增加词汇-部首注意力偏置,从而增强模型的识别能力。在自建的数据集上进行多组对比实验,结果表明,在力学问题实体识别任务中,相对于WCGCN算法,该算法在精度、召回率和F1值上分别提高1.92、0.99和1.44个百分点,能够有效提取力学问题中的关键信息。

关键词: 命名实体识别, 多元信息嵌入, 图注意力网络, Transformer架构, 注意力机制

Abstract:

Automatic extraction of key entities is important for solving mechanical problems automatically. However, compared to open domains, mechanical problems entail a larger specialized vocabulary and longer entities, making their recognition difficult and thus lowering the accuracy. To address this issue, an entity recognition algorithm based on a Graph Attention Network (GAT) and Transformer encoder is proposed for key entity extraction in mechanical problems. A Multi-Meta Information Embedding (MMIE) that includes character information, lexical information, and radical information is designed to enhance Chinese sentence representation based on the characteristics of Chinese. Additionally, two graph models, Structural Graph (S-Graph) and Contextual Graph (C-Graph), are proposed to model Chinese sentences. To learn the contextual information of sentences, a collaborative architecture is designed that integrates multiple types of information using two independent GATs. To balance the impact of lexical information and radical information on Chinese characters, a collaborative Transformer architecture is proposed, consisting of a character-lexical Transformer and a character-radical Transformer. A lexical-radical attention bias is added to enhance the recognition ability of the model. Several comparative experiments were conducted on a self-built dataset, and the results show that the proposed algorithm improves precision, recall, and F1 values by 1.92, 0.99, and 1.44 percentage points compared to Word-Character Graph Convolution Network(WCGCN) algorithm, respectively, for the entity recognition task in mechanical problems, and the algorithm effectively extracts key information in mechanical problems.

Key words: Named Entity Recognition(NER), Multi-Meta Information Embedding(MMIE), Graph Attention Network(GAT), Transformer architecture, attention mechanism