期刊文献+

基于激活-熵的分层迭代剪枝策略的CNN模型压缩 被引量:2

CNN model compression based on activation-entropy based layer-wise iterative pruning strategy
在线阅读 下载PDF
导出
摘要 针对卷积神经网络(CNN)模型现有剪枝策略各尽不同和效果一般的情况,提出了基于激活-熵的分层迭代剪枝(AE-LIP)策略,保证模型精度在可控范围内的同时缩减模型的参数量。首先,结合神经元激活值和信息熵,构建基于激活-熵的权重评判准则,计算权值重要性得分;然后,逐层剪枝,根据重要性得分对权值排序,并结合各层剪枝数量筛选出待剪枝权重并将其设置为0;最后,微调模型,重复上述过程,直至迭代结束。实验结果表明,采用基于激活-熵的分层迭代剪枝策略:AlexNet模型压缩了87.5%;相应的准确率下降了2.12个百分点,比采用基于幅度的权重剪枝策略提高了1.54个百分点,比采用基于相关性的权重剪枝策略提高0.91个百分点。VGG-16模型压缩了84.1%;相应的准确率下降了2.62个百分点,比采用上述两个对比策略分别提高了0.62个百分点和0.27个百分点。说明所提策略在保证模型精确度下有效缩减了CNN模型的大小,有助于CNN模型在存储受限的移动设备上的部署。 Since the existing pruning strategies of the Convolutional Neural Network(CNN)model are different and have general effects,an Activation-Entropy based Layer-wise Iterative Pruning(AE-LIP)strategy was proposed to reduce the parameter amount of the model while ensuring the accuracy of the model within a controllable range.Firstly,combined with the neuronal activation value and information entropy,a weight evaluation criteria based on activation-entropy was constructed,and the weight importance score was calculated.Secondly,the pruning was performed layer by layer,the weights were sorted according to the importance score,and the pruning number in each layer was combined to filter out the weights to be pruned and set them to zero.Finally,the model was fine-tuned,and the above process was repeated until the iteration ended.The experimental results show that the activation-entropy based layer-wise iterative pruning strategy makes the AlexNet model compressed 87.5%,and the corresponding accuracy is reduced by 2.12 percentage points,which is 1.54 percentage points higher than that of the magnitude-based weight pruning strategy and 0.91 percentage points higher than that of the correlation-based weight pruning strategy;the strategy makes VGG-16 model compressed 84.1%,and the corresponding accuracy is reduced by 2.62 percentage points,which is 0.62 and 0.27 percentage points higher than those of the two above strategies.It can be seen that the proposed strategy reduces the size of the CNN model effectively while ensuring the accuracy of the model,and is helpful for the deployment of CNN model on mobile devices with limited storage.
作者 陈程军 毛莺池 王绎超 CHEN Chengjun;MAO Yingchi;WANG Yichao(College of Computer and Information,Hohai University,Nanjing Jiangsu 211100,China)
出处 《计算机应用》 CSCD 北大核心 2020年第5期1260-1265,共6页 journal of Computer Applications
基金 “十三五”国家重点研发计划项目(2018YFC0407105) 华能集团重点研发项目(HNKJ17-21)。
关键词 移动云计算 神经元激活值 信息熵 迭代剪枝 模型压缩 mobile cloud computing neuronal activation value information entropy iterative pruning model compression
  • 相关文献

参考文献2

二级参考文献2

共引文献39

同被引文献18

引证文献2

二级引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部