作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2021, Vol. 47 ›› Issue (5): 80-87. doi: 10.19678/j.issn.1000-3428.0057842

• 人工智能与模式识别 • 上一篇    下一篇

结合半波高斯量化与交替更新的神经网络压缩方法

张红梅, 严海兵, 张向利   

  1. 桂林电子科技大学 广西高校云计算与复杂系统重点实验室, 广西 桂林 541004
  • 收稿日期:2020-03-24 修回日期:2020-04-26 发布日期:2020-05-13
  • 作者简介:张红梅(1970-),女,教授、博士,主研方向为网络信息安全、嵌入式系统、智能信息处理;严海兵,硕士研究生;张向利,教授、博士。
  • 基金资助:
    国家自然科学基金(61461010);认知无线电与信息处理省部共建教育部重点实验室基金(CRKL170103,CRKL170104);广西密码学与信息安全重点实验室基金(GCIS201626)。

Neural Network Compression Method Combining Half-Wave Gaussian Quantization and Alternate Update

ZHANG Hongmei, YAN Haibing, ZHANG Xiangli   

  1. Guangxi Colleges and Universities Key Laboratory of Cloud Computing and Complex Systems, Guilin University of Electronic Technology, Guilin, Guangxi 541004, China
  • Received:2020-03-24 Revised:2020-04-26 Published:2020-05-13

摘要: 为使神经网络模型能在实时性要求较高且内存容量受限的边缘设备上部署使用,提出一种基于半波高斯量化与交替更新的混合压缩方法。对神经网络模型输入部分进行2 bit均匀半波高斯量化,将量化值输入带有缩放因子的二值网络通过训练得到初始二值模型,利用交替更新方法对已训练的二值模型进行逐层微调以提高模型测试精度。在CIFAR-10和ImageNet数据集上的实验结果表明,该方法能有效降低参数和结构冗余所导致的内存和时间开销,在神经网络模型压缩比接近30的前提下,测试精度相比HWGQ-Net方法提高0.8和2.0个百分点且实现了10倍的训练加速。

关键词: 卷积神经网络, 量化, 模型压缩, 半波高斯量化, 交替更新

Abstract: To enable the deployment of neural network models on edge devices with a limited memory size and high real-time performance requirements,this paper proposes a hybrid compression method combining Half-Wave Gaussian Quantization(HWGQ) and alternate update.By performing the 2 bit uniform HWGQ on the input of the neural network model,the quantized value is input into a binary network with a scaling factor,which is trained to obtain the initial binary model.Then the trained binary model is fine-tuned layer by layer using the alternating update method to improve the accuracy of the model.Experimental results on the CIFAR-10 and ImageNet datasets show that the proposed method significantly reduces the memory consumption and time consumption caused by parameter redundancy and structural redundancy.When the model compression ratio is about 30,the accuracy of the model is increased by 0.8 and 2.0 percentage points compared with that of the HWGQ-Net method,and its training speed is increased by 10 times.

Key words: Convolutional Neural Network(CNN), quantization, model compression, Half-Wave Gaussian Quantization(HWGQ), alternate update

中图分类号: