Design of Computing-in-Memory (CIM) with Vertical Split-Gate Flash Memory for Deep Neural Network (DNN) Inference Accelerator

2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS)(2021)

引用 9|浏览4
暂无评分
摘要
Computing-In-Memory (CIM) using Flash memory is a potential solution to support a heavy-weight DNN inference accelerator for edge computing applications. Flash memory provides the best high-density and low-cost non-volatile memory solution to store the weights, while CIM functions of Flash memory can compute AI neural network calculations inside the memory chip. Our analysis indicates that Flash CIM can save data movements by similar to 85% as compared with the conventional Von-Neumann architecture. In this work, we propose a detail device and design co-optimizations to realize Flash CIM, using a novel vertical split-gate Flash device. Our device supports low-voltage (<1V) read at WL's and BL's, tight and tunable cell current (Icell) ranging from 150nA to 1.5uA, extremely large Icell ON/OFF ratio similar to 7 orders, small RTN noise and negligible read disturb to provide a high-performance and highly-reliable CIM solution.
更多
查看译文
关键词
design co-optimizations,low-cost nonvolatile memory solution,low-voltage read,conventional von-Neumann architecture,data movements,novel vertical split-gate flash device,highly-reliable CIM solution,Flash CIM,memory chip,AI neural network calculations,CIM functions,edge computing applications,heavy-weight DNN inference accelerator,deep neural network inference accelerator,vertical split-gate flash memory,computing-in-memory design,voltage 1.0 V,current 150.0 nA
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要