[1] 陆伟,武川.实体链接研究综述[J].情报学报,2015,34(1):105-112. LU W,WU C.Literature review on entity[J].Journal of the China Society for Scientific and Technical Information,2015,34(1):105-112.(in Chinese) [2] 徐馨韬,柴小丽,谢彬,等.基于改进TextRank算法的中文文本摘要提取[J].计算机工程,2019,45(3):273-277. XU X T,CHAI X L,XIE B,et al.Extraction of Chinese text summarization based on improved TextRank algorithm[J]. Computer Engineering,2019,45(3):273-277.(in Chinese) [3] 郭宇航,秦兵,刘挺,等.实体链指技术研究进展[J].智能计算机与应用,2014,4(5):9-13. GUO Y H,QIN B,LIU T,et al.Research progress of entity linking[J].Intelligent Computer and Applications,2014,4(5):9-13.(in Chinese) [4] FINKEL J R,GRENAGER T,MANNING C.Incorporating non-local information into information extraction systems by gibbs sampling[C]//Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics.[S.l.]:ACL,2005:363-370. [5] HE J Z,WANG H F.Chinese named entity recognition and word segmentation based on character[C]//Proceedings of the 6th SIGHAN Workshop on Chinese Language Processing.[S.l.]:ACL,2008:128-132. [6] MAO X N,HE S K,BAO S C,et al.Chinese word segmentation and named entity recognition based on conditional random fields[C]//Proceedings of the 6th SIGHAN Workshop on Chinese Language Processing.[S.l.]:ACL,2007:90-93. [7] LAMPLE G,BALLESTEROS M,SUBRAMANIAN S,et al.Neural architectures for named entity recognition[EB/OL].(2016-04-07)[2020-09-10].https://arxiv.org/pdf/1603.01360v3.pdf. [8] TAN Z X,WANG M X,XIE J,et al.Deep semantic role labeling with self-attention[C]//Proceedings of the 32nd AAAI Conference on Artificial Intelligence.[S.l.]:AAAI Press,2018:1-5. [9] 李天然,刘明童,张玉洁,等.基于深度学习的实体链接方法研究[J].北京大学学报(自然科学版),2021,57(1):91-98. LI T R,LIU M T,ZHANG Y J,et al.A review of entity linking research based on deep learning[J].Acta Scientiarum Naturalium Universitatis Pekinensis,2021,57(1):91-98. [10] SHEN W,WANG J Y,LUO P,et al.LINDEN:linking named entities with knowledge base via semantic knowledge[C]//Proceedings of Annual Conference on World Wide Web.Lyon,France:[s.n.]:2012:1-5. [11] GANEA O E,HOFMANN T.Deep joint entity disambiguation with local neural attention[EB/OL].(2017-08-29)[2020-09-10].https://arxiv.org/pdf/1704.04920.pdf. [12] KOLITSAS N,GANEA O E,HOFMANN T.End-to-end neural entity linking[EB/OL].(2018-08-29)[2020-09-10].https://arxiv.org/pdf/1808.07699v2.pdf. [13] CAO P F,CHEN Y B,LIU K,et al.Adversarial transfer learning for Chinese named entity recognition with self-attention mechanism[C]//Proceedings of 2018 Conference on Empirical Methods in Natural Language Processing.Brussels,Belgium:[s.n.],2018:1-5. [14] SIL A,YATES A.Re-ranking for joint named-entity recognition and linking[C]//Proceedings of the 22nd ACM International Conference on Information and Knowledge Management.New York,USA:ACM Press,2013:2369-2374. [15] 怀宝兴,宝腾飞,祝恒书,等.一种基于概率主题模型的命名实体链接方法[J].软件学报,2014,25(9):2076-2087. HUAI B X,BAO T F,ZHU H S,et al.Topic modeling approach to named entity linking[J].Journal of Software,2014,25(9):2076-2087.(in Chinese) [16] LUO G,HUANG X J,LIN C Y,et al.Joint entity recognition and disambiguation[C]//Proceedings of 2016 Conference on Empirical Methods in Natural Language Processing.Austin,USA:[s.n.],2016:1-5. [17] YANG X Y,GU X T,LIN S,et al.Learning dynamic context augmentation for global entity linking[EB/OL].(2019-09-04)[2020-09-10].https://arxiv.org/pdf/1909. 02117.pdf. [18] XUE M G,CAI W M,SU J S,et al.Neural collective entity linking based on recurrent random walk network learning[EB/OL].(2019-06-20)[2020-09-10].https://arxiv.org/pdf/1906.09320.pdf. [19] 吴晓崇,段跃兴,张月琴,等.基于CNN和深层语义匹配的中文实体链接模型[J].计算机工程与科学,2020,42(8):1514-1520. WU X C,DUAN Y X,ZHANG Y Q,et al.A Chinese entity linking model based on CNN and deep structured sematic model[J].Computer Engineering & Science,2020,42(8):1514-1520.(in Chinese) [20] PETERS M,NEUMANN M,IYYER M,et al.Deep contextualized word representations[C]//Proceedings of 2018 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.[S.l.]:ACL,2018:2227-2237. [21] VASWANI A,SHAZEER N,PARMAR N,et al.Attention is all you need[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems.New York,USA:ACM Press,2017:5998-6008. [22] DEVLIN J,CHANG M W,LEE K,et al.BERT:pre-training of deep bidirectional transformers for language understanding[EB/OL].(2018-10-11)[2020-09-10].https://arxiv.org/pdf/1810.04805v1.pdf. [23] 詹飞,朱艳辉,梁文桐,等.基于BERT和TextRank关键词提取的实体链接方法[J].湖南工业大学学报,2020,34(4):63-70. ZHAN F,ZHU Y H,LIANG W T,et al.Entity linking via BERT and TextRank keyword extraction[J].Journal of Hunan University of Technology,2020,34(4):63-70.(in Chinese) [24] LUONG M T,PHAM H,MANNING C D.Effective approaches to attention-based neural machine translation[EB/OL].(2015-09-20)[2020-09-10].https://arxiv.org/pdf/1508.04025.pdf. |