Author Login Editor-in-Chief Peer Review Editor Work Office Work

Computer Engineering ›› 2022, Vol. 48 ›› Issue (11): 62-68,76. doi: 10.19678/j.issn.1000-3428.0063185

• Artificial Intelligence and Pattern Recognition • Previous Articles     Next Articles

Multi-Modal Fine-Grained Retrieval Based on Modal Specific and Modal Shared Feature Information

LI Pei1,2, CHEN Qiaosong1,2, CHEN Pengchang1,2, DENG Xin1,2, WANG Jin1,2, PIAO Changhao1,2   

  1. 1. College of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China;
    2. Chongqing Key Laboratory of Data Engineering and Visual Computing, Chongqing 400065, China
  • Received:2021-11-09 Revised:2021-12-29 Published:2022-01-12

基于模态特异及模态共享特征信息的多模态细粒度检索

李佩1,2, 陈乔松1,2, 陈鹏昌1,2, 邓欣1,2, 王进1,2, 朴昌浩1,2   

  1. 1. 重庆邮电大学 计算机科学与技术学院, 重庆 400065;
    2. 数据工程与认知计算重庆市重点实验室, 重庆 400065
  • 作者简介:李佩(1996—),男,硕士研究生,主研方向为多模态检索、机器视觉;陈乔松,副教授、博士;陈鹏昌,硕士;邓欣,副教授、博士;王进,教授、博士;朴昌浩,教授、博士后。
  • 基金资助:
    国家自然科学基金(61806033);国家社会科学基金西部项目(18XGL013)。

Abstract: The goal of cross-modal retrieval is that the user gives any sample as a query sample;then, the system retrieves and feeds back various modal samples related to the query sample.Multi-modal fine-grained retrieval emphasizes that the number of modalities is greater than two and the granularity of classification is the fine-grained sub-category.This paper introduces the concepts of modal specific features and modal shared features and proposes the MS2Net framework.The branch network and backbone network are used to extract the modal specific features and modal shared features of different modal data.Then, the two features are fully fused through the Multi-Modal Feature fusion Module(MMFM).Meanwhile, the semantic information contained in the high-dimensional space vector is greatly increased by using the unique information of each mode and the commonness and relationship between different modal data.In addition, for the multi-modal fine-grained retrieval scenario, based on center loss, this paper proposes multi-center loss, introduces the inner-class center to gather the samples of the same category and the same mode, and then indirectly gathers the samples of the same category but different modes by aggregating the inner-class center.This reduces the heterogeneous gap and semantic gap between the samples.It clearly enhances the clustering ability of the model to high-dimensional spatial vectors.Finally, the experimental results of one-to-one and one-to-multimodal retrieval on the FG-Xmedia public dataset show that, compared with the FGCrossNet method, the MS2Net method improves the mAP index by 65% and 48%, respectively.

Key words: information retrieval, multi-modal retrieval, fine-grained retrieval, multi-modal representation learning, deep learning

摘要: 跨模态检索的目标是用户给定任意一个样本作为查询样例,系统检索得到与查询样例相关的各个模态样本,多模态细粒度检索在跨模态检索基础上强调模态的数量至少大于两个,且待检索样本的分类标准为细粒度子类,存在多模态数据间的异构鸿沟及细粒度样本特征差异小等难题。引入模态特异特征及模态共享特征的概念,提出一种多模态细粒度检索框架MS2Net。使用分支网络及主干网络分别提取不同模态数据的模态特异特征及模态共享特征,将两种特征通过多模态特征融合模块进行充分融合,同时利用各个模态自身的特有信息及不同模态数据间的共性及联系,增加高维空间向量中包含的语义信息。针对多模态细粒度检索场景,在center loss函数的基础上提出multi-center loss函数,并引入类内中心来聚集同类别且同模态的样本,根据聚集类内中心来间接聚集同类别但模态不同的样本,同时消减样本间的异构鸿沟及语义鸿沟,增强模型对高维空间向量的聚类能力。在公开数据集FG-Xmedia上进行一对一与一对多的模态检索实验,结果表明,与FGCrossNet方法相比,MS2Net方法mAP指标分别提升65%和48%。

关键词: 信息检索, 多模态检索, 细粒度检索, 多模态表征学习, 深度学习

CLC Number: