On Continuing DNN Accelerator Architecture Scaling Using Tightly Coupled Compute-on-Memory 3-D ICs

IEEE Transactions on Very Large Scale Integration (VLSI) Systems(2023)

引用 0|浏览13
暂无评分
摘要
This work identifies the architectural and design scaling limits of 2-D flexible interconnect deep neural network (DNN) accelerators and addresses them with 3-D ICs. We demonstrate how scaling up a baseline 2-D accelerator in the $X/Y$ dimension fails and how vertical stacking effectively overcomes the failure. We designed multitier accelerators that are $1.67\times $ faster than the 2-D design. Using our 3-D architecture and circuit codesign methodology, we improve throughput, energy efficiency, and area efficiency by up to $5\times $ , $1.2\times $ , and $3.9\times $ , respectively, over 2-D counterparts. The IR-drop in our 3-D designs is within 10.7% of VDD, and the temperature variation is within 12 °C.
更多
查看译文
关键词
dnn accelerator architecture scaling,compute-on-memory
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要