Pcnn: Pattern-Based Fine-Grained Regular Pruning Towards Optimizing Cnn Accelerators

PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC)(2020)

引用 24|浏览78
暂无评分
摘要
Weight pruning is a powerful technique to realize model compression. We propose PCNN, a fine-grained regular 1D pruning method. A novel index format called Sparsity Pattern Mask (SPM) is presented to encode the sparsity in PCNN. Leveraging SPM with limited pruning patterns and non-zero sequences with equal length, PCNN can be efficiently employed in hardware. Evaluated on VGG-16 and ResNet-18, our PCNN achieves the compression rate up to 8.4 x with only 0.2% accuracy loss. We also implement a pattern-aware architecture in 55nm process, achieving up to 9.0x speedup and 28.39 TOPS/VV efficiency with only 3.1% on-chip memory overhead of indices.
更多
查看译文
关键词
optimizing pcnn,pattern-based,fine-grained
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要