文献阅读(97)网络压缩综述(部分)

  • 题目:Deep Neural Network Approximation for Custom Hardware: Where We’ve Been, Where We’re Going
  • 时间:2019
  • 期刊:ACM Computing Survey
  • 研究机构:伦敦帝国学院

1 缩写 & 引用

2 abstract & introduction

本片综述主要总结了:

  • 高性能神经网络在硬件上的近似方法
  • 比较了不同硬件平台的屋顶模型
  • 评估了每种方法的性能
  • 未来发展的方向

DNN近似算法主要分成两类:量化和权重减枝
性能评估指标:

  • 精度
  • 压缩率
  • Thoughput: Classifications processing per second
  • 延时
  • 能耗比:单位功耗的throughput

3 量化

3.1 fixed-point

blocking floating point(BFP)=dynamic fixed point
[111] Going deeper with embedded FPGA platform for convolutional neural network 2016 FPGA

formulat an optimisation problem for minimising quantisation error with respect to changes in precision and binary point location

[128]Fixed-point optimization of deep neural networks with adaptive step size retraining

treating quantisation resolution as a trainable parameter

[69] Adaptive quantization of neural networks 2018 ICLR

investigated quantisation at a finer granularity than the aforementioned down-to layer-wise methods
During retraining, networks adapt, with each filter allowed to assume an independent precision.Experiments with small-scale datasets and models showed that Adaptive Quantisation, when combined with pruning,is able to achieve accuracies and compression ratios superior to binarised neural networks

[161] DoReFa-Net: Training low bitwidth convolutional neural networks with low bitwidth gradients 2016 arXiv

DoReFa-Net, supports arbitrary precisions for weights, activations, and gradients, from 32-bit fixed point down
to binary

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值