作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2012, Vol. 38 ›› Issue (22): 167-170. doi: 10.3969/j.issn.1000-3428.2012.22.041

• 人工智能及识别技术 • 上一篇    下一篇

基于粒计算的K近邻多标签学习算法

陈小波,吴 涛,高正龙   

  1. (安徽大学数学科学学院,合肥 230039)
  • 收稿日期:2012-01-05 修回日期:2012-03-15 出版日期:2012-11-20 发布日期:2012-11-17
  • 作者简介:陈小波(1985-),男,硕士,主研方向:智能计算;吴 涛,教授、博士;高正龙,硕士
  • 基金资助:

    国家自然科学基金资助项目(61073117);国家“973”计划基金资助项目(2007BC311003);安徽大学学术创新团队基金资助项目(KJTD001B);安徽大学研究生学术创新基金资助项目(yfc090008)

K-nearest Neighbor Multi-label Learning Algorithm Based on Granular Computing

CHEN Xiao-bo, WU Tao, GAO Zheng-long   

  1. (School of Mathematical Sciences, Anhui University, Hefei 230039, China)
  • Received:2012-01-05 Revised:2012-03-15 Online:2012-11-20 Published:2012-11-17

摘要: K近邻多标签学习算法的近邻点个数取固定值,而没有考虑样本分布的特点,可能会将相似度高的点排除在近邻集外,或者将相似度低的点包含在近邻集内,影响分类器的性能。为此,将粒计算的思想引入近邻集的构建,提出一种新的K近邻多标签学习算法。通过粒度控制,确定近邻点集,使得领域内的样本点有高相似性,且此类样本能进入近邻集。实验结果表明,该算法的大多数评价指标均优于现有的多标签学习算法。

关键词: 多标签学习, 粒计算, K近邻, 粒度, 评价指标

Abstract: In Multi-label K-nearest Neighbor(ML-KNN) learning algorithm, the number of nearest neighbors is given in prior and its value is chosen without considering the distribution of samples, it is possible that highly similar samples are not in the nearest neighbor or low similar samples are in the nearest neighbor set, which affect the performance of the classifier. In view of this case, a novel ML-KNN algorithm is put forward based on the idea of Granular Computing(GrC), the nearest neighbor set is constructed with the controlling of the granular hierarchy, and the nearest neighbors of a sample have high similarity and highly similar samples can be added to nearest neighbor set. Experimental results show that most of the evaluation criteria in new algorithm are better than the traditional algorithm.

Key words: multi-label learning, Granular Computing(GrC), K-nearest Neighbor(KNN), granularity, evaluation index

中图分类号: