On-Chip Memory Access Reduction for Energy-Efficient Dilated Convolution Processing.

Simon Friedrich, Thomas Nalapat,Robert Wittig,Emil Matús,Gerhard P. Fettweis

SAMOS(2023)

引用 0|浏览1
暂无评分
摘要
Dilated convolutions have recently become increasingly popular in deep neural networks. However, the inference of these operations on hardware accelerators is not mature enough to reach the efficiency of standard convolutions. Therefore, we extended a dedicated accelerator for dilated convolutions to reduce the number of energy-intensive accesses to the on-chip memory. We achieve this by applying the principle of feature map decomposition to an output-stationary compute array with a strided feature loading. Our solution shows a 50% reduction in memory accesses for an unpadded 3 $$\,\times \,$$ 3 kernel and a dilation rate of 9 compared to a recently proposed dilated convolution accelerator. We also support flexible parameter selection for kernel sizes and dilation rates to meet the requirements of modern neural networks. The energy consumption of the additional hardware modules is less than the savings achieved by the reduced memory accesses. This results in a relative energy saving by a factor of 4.77 for dilated convolutions with unpadded 3 $$\,\times \,$$ 3 kernels.
更多
查看译文
关键词
memory,convolution,on-chip,energy-efficient
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要