作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2022, Vol. 48 ›› Issue (5): 67-73. doi: 10.19678/j.issn.1000-3428.0061461

• 人工智能与模式识别 • 上一篇    下一篇

融合弱层惩罚的卷积神经网络模型剪枝方法

房志远, 石守东, 郑佳罄, 胡加钿   

  1. 宁波大学 信息科学与工程学院, 浙江 宁波 315211
  • 收稿日期:2021-04-26 修回日期:2021-06-19 发布日期:2021-05-25
  • 作者简介:房志远(1995—),男,硕士研究生,主研方向为模型压缩与加速;石守东,副教授、博士;郑佳罄、胡加钿,硕士研究生。
  • 基金资助:
    宁波市公益项目“基于深度学习的儿童学习姿态识别系统研究与实现”(2019C50020)。

Pruning Method of Convolutional Neural Network Model with Weak Layer Penalty

FANG Zhiyuan, SHI Shoudong, ZHENG Jiaqing, HU Jiadian   

  1. Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, Zhejiang 315211, China
  • Received:2021-04-26 Revised:2021-06-19 Published:2021-05-25

摘要: 深度卷积神经网络的存储和计算需求巨大,难以在一些资源受限的嵌入式设备上进行部署。为尽可能减少深度卷积神经网络模型在推理过程中的资源消耗,引入基于几何中值的卷积核重要性判断标准,提出一种融合弱层惩罚的结构化非均匀卷积神经网络模型剪枝方法。使用欧式距离计算各层卷积核间的信息距离,利用各卷积层信息距离的数据分布特征识别弱层,通过基于贡献度的归一化函数进行弱层惩罚,消除各层间的差异性。在全局层面评估卷积核重要性,利用全局掩码技术对所有卷积核实现动态剪枝。在CIFAR-10、CIFAR-100和SVHN数据集上的实验结果表明,与SFP、PFEC、FPGM和MIL剪枝方法相比,该方法剪枝得到的VGG16单分支、Resnet多分支、Mobilenet-v1轻量化网络模型在保证精度损失较小的情况下,有效地减少了模型参数量和浮点操作数。

关键词: 模型剪枝, 弱层惩罚, 全局掩码, 欧式距离, 核重要性评估

Abstract: The extensive demand of convolutional neural networks for memory and computation makes it difficult to deploy them in resource-constrained embedded devices.To minimize the resource consumption of the deep convolutional neural network model during the inference process, this study introduces a criterion for judging the importance of the convolution kernel, based on the geometric median and further proposes a structured, non-uniform pruning method for convolutional neural network models with weak layer penalty.First, using the Euclidean distance, the algorithm calculates the information distance for each layer of the convolution kernel.Then, the data distribution characteristics of the information distance of each convolutional layer are used to identify the weak layers, and a normalization function based on the contribution degree is proposed to eliminate the difference between layers while weakening the redundant layers.Second, the importance of the convolution kernel is evaluated at the global level, and the global mask technique is used to achieve dynamic pruning.The experimental results on the CIFAR-10, CIFAR-100, and SVHN datasets demonstrate that compared with SFP, PFEC, FPGM, and MIL pruning methods, the proposed method prunes the VGG16 single-branch, Resnet multi-branch, and Mobilenet-v1 lightweight network models, effectively reducing the number of model parameters and Floating Points of Operations(FLOPs) while ensuring that the loss of precision is small.

Key words: model pruning, weak layer penalty, global mask, Euclidean distance, kernel importance evaluation

中图分类号: