深度神经网络加速和压缩

本文总结了神经网络加速和压缩的方法,包括低秩分解、剪枝、量化、知识蒸馏和紧凑网络设计。低秩方法如SVD、CP分解等已较少使用,剪枝技术如结构化剪枝、网络瘦身、梯度剪枝得到发展。量化技术通过低比特量化、二值权重网络等降低计算复杂度。知识蒸馏则通过定义知识和优化损失来传递教师网络的知识。紧凑网络设计如MobileNet和ShuffleNet在保持效率的同时减少了计算量。
摘要由CSDN通过智能技术生成

模型加速与压缩方法分类总结

• Low-Rank
• Pruning
• Quantization
• Knowledge Distillation
• Compact Network Design

 

Low-Rank

Previous low-rank based methods:


• SVD
- Zhang et al., “Accelerating Very Deep Convolutional Networks for Classification and Detection”. IEEE TPAMI 2016.


• CP decomposition
- Lebedev et al., “Speeding-up Convolutional Neural Networks Using Fine-tuned CP- Decomposition”. ICLR 2015.


• Tucker decomposition
- Kim et al., “Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications”. ICLR 2016.


• Tensor Train Decomposition
- Novikov et al., “Tensorizing Neural Networks”. NIPS 2016.


• Block Term Decomposition
- Wang et al., “Accelerating Convolutional Neural Networks for Mobile Applications”. ACMMM 2016.

 

Recent low-rank based methods:


• Tensor Ring (TR) factorizations
- Wang et al., “Wide Compression: Tensor Ring Nets”. CVPR2018


• Block Term Decomposition For RNN
- Ye et al., “Learning Compact Recurrent Neural Networks with Block-Term Tensor Decomposition ”. CVPR2018.


Why low-rank is not popular anymore?


• Low-rank approximation is not efficient for those 1x1 convolutions
• 3x3 convolutions in bottleneck structure have less computation complexity
• Depthwise convolution or grouped 1x1 convolution is already quite fast.

 

Pruning

Recent progress in pruning :

• Structured Pruning
– Yoon et al. “Combined Group and Exclusive Sparsity for Deep Neural Networks”. ICML2017
– Ren et al. “SBNet: Sparse Blocks Network for Fast Inference”. CVPR2018


• Filter Pruning
– Luo et al. “Thinet: A filter level pruning method for deep neural network compression”. ICCV2017
– Liu et al., “Learning efficient convolutional networks through network slimming”. ICCV2017
– He et al. “Channel Pruning for Accelerating Very Deep Neural Networks”. ICCV2017


• Gradient Pruning
– Sun et al. “meProp: Sparsified Back Propagation for Accelerated Deep Learning with Reduced Overfitting”. ICML2017


• Fine-grained Pruning in a Bayesian View
– Molchanov et al. “Variational Dropout Sparsifies Deep Neural Networks”. ICML2017

Structured Pruning

Previous group pruning methods mainly use the group sparsity,Yoon et al. use both group sparsity and exclusive sparsity

• Group Sparsity: Impose sparsity regularization on grouped features to prune columns of weight matrix.

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值