作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2023, Vol. 49 ›› Issue (12): 96-102. doi: 10.19678/j.issn.1000-3428.0066449

• 人工智能与模式识别 • 上一篇    下一篇

基于全局节点和多片段的格栅命名实体识别

郭江涛, 彭甫镕*   

  1. 山西大学 大数据科学与产业研究院, 太原 030006
  • 收稿日期:2022-12-05 出版日期:2023-12-15 发布日期:2023-12-14
  • 通讯作者: 彭甫镕
  • 作者简介:

    郭江涛(1996—),男,硕士,主研方向为自然语言处理

  • 基金资助:
    国家自然科学基金面上项目(62276162)

Lattice Named Entity Recognition Based on Global Nodes and Multi-fragments

Jiangtao GUO, Furong PENG*   

  1. Institute of Big Data Science and Industry, Shanxi University, Taiyuan 030006, China
  • Received:2022-12-05 Online:2023-12-15 Published:2023-12-14
  • Contact: Furong PENG

摘要:

现有命名实体识别模型对标注数据量要求较高,基于主动学习的命名实体识别模型需要人工分词造成标注代价大。针对上述问题,提出一种结合全局节点和多片段的格栅命名实体识别模型。将Transformer的全连接结构替换为全局节点和多片段结构,每个节点仅与构造的上下文向量进行注意力计算,全局和片段节点分别获取全局和局部信息,从而降低对标注数据的需求量。对Flat-Lattice结构进行改进,解决现有主动学习策略需要分词的问题,从而在保证模型性能的情况下降低数据标注代价。在MSRA、OntoNotes 5.0、Weibo、PeopleDaily这4个数据集上的实验结果表明,与FLAT模型相比,所提模型达到对应F1阈值所需的标注数据量分别降低了39.90%、2.17%、34.60%和35.67%。

关键词: 全局节点, 多片段, 格栅, 命名实体识别, 主动学习, Transformer结构

Abstract:

A lattice Named Entity Recognition(NER) model that combines global nodes and multiple fragments is proposed to address the high demand for annotation data in existing NER models and the high cost of annotation incurred by manual word segmentation in active learning-based NER models. The fully connected structure of the Transformer is replaced with a global node and multi-fragment structure, where each node performs only attention calculations with the constructed context vector. The global and fragmented nodes obtain global and local information separately, thereby reducing the demand for annotated data. The Flat-Lattice structure is improved to solve the problem of word segmentation in existing active learning strategies, thereby reducing the cost of data annotation and ensuring model performance. The experimental results on four datasets, namely MSRA, OntoNotes 5.0, Weibo, and PeopleDaily, show that compared with the FLAT model, the proposed model reduces the amount of annotated data required to reach the corresponding F1 threshold by 39.90%, 2.17%, 34.60%, and 35.67%, respectively.

Key words: global node, multi-fragment, lattice, Named Entity Recognition(NER), active learning, Transformer structure