文章目录
- 题目:UNPU: A 50.6TOPS/W Unified Deep Neural Network Accelerator with 1b-to-16b Fully-Variable Weight Bit-Precision
- 时间:2018
- 会议:ISSCC(13.3)
- 研究机构:KAIST韩国科学技术院
1 英文缩写
- CL: convolutional layer
- FCL: fully-connected layer
- RL: recurrent layer
- PE: processing element
- UNPU: unified neural processing unit
- IF: input feature
- LBPE: lookup-table-based bit-serial PE
- AFL: aligned feature loader
- OF: output feature
2 overall architecture
In this paper, we present a unified neural processing unit (UNPU) supporting CLs, RLs, and FCLs with fully-variable weight bit-precision from 1b to 16b.
- reuse of input feature
- the lookup-table-based bit serial PE is implement for energy-optimal DNN operations with variable-weight bit-precisions from 1b to 1