Author Login Editor-in-Chief Peer Review Editor Work Office Work

Computer Engineering ›› 2012, Vol. 38 ›› Issue (22): 167-170. doi: 10.3969/j.issn.1000-3428.2012.22.041

• Networks and Communications • Previous Articles     Next Articles

K-nearest Neighbor Multi-label Learning Algorithm Based on Granular Computing

CHEN Xiao-bo, WU Tao, GAO Zheng-long   

  1. (School of Mathematical Sciences, Anhui University, Hefei 230039, China)
  • Received:2012-01-05 Revised:2012-03-15 Online:2012-11-20 Published:2012-11-17

基于粒计算的K近邻多标签学习算法

陈小波,吴 涛,高正龙   

  1. (安徽大学数学科学学院,合肥 230039)
  • 作者简介:陈小波(1985-),男,硕士,主研方向:智能计算;吴 涛,教授、博士;高正龙,硕士
  • 基金资助:

    国家自然科学基金资助项目(61073117);国家“973”计划基金资助项目(2007BC311003);安徽大学学术创新团队基金资助项目(KJTD001B);安徽大学研究生学术创新基金资助项目(yfc090008)

Abstract: In Multi-label K-nearest Neighbor(ML-KNN) learning algorithm, the number of nearest neighbors is given in prior and its value is chosen without considering the distribution of samples, it is possible that highly similar samples are not in the nearest neighbor or low similar samples are in the nearest neighbor set, which affect the performance of the classifier. In view of this case, a novel ML-KNN algorithm is put forward based on the idea of Granular Computing(GrC), the nearest neighbor set is constructed with the controlling of the granular hierarchy, and the nearest neighbors of a sample have high similarity and highly similar samples can be added to nearest neighbor set. Experimental results show that most of the evaluation criteria in new algorithm are better than the traditional algorithm.

Key words: multi-label learning, Granular Computing(GrC), K-nearest Neighbor(KNN), granularity, evaluation index

摘要: K近邻多标签学习算法的近邻点个数取固定值,而没有考虑样本分布的特点,可能会将相似度高的点排除在近邻集外,或者将相似度低的点包含在近邻集内,影响分类器的性能。为此,将粒计算的思想引入近邻集的构建,提出一种新的K近邻多标签学习算法。通过粒度控制,确定近邻点集,使得领域内的样本点有高相似性,且此类样本能进入近邻集。实验结果表明,该算法的大多数评价指标均优于现有的多标签学习算法。

关键词: 多标签学习, 粒计算, K近邻, 粒度, 评价指标

CLC Number: