An Architecture of Sparse Length Sum Accelerator in AxDIMM

2022 IEEE 4th International Conference on Artificial Intelligence Circuits and Systems (AICAS)(2022)

引用 1|浏览18
暂无评分
摘要
In this paper, we have implemented high-efficient near-memory sparse-length sum hardware accelerator which is parallelized over each channel or rank to support Meta's deep learning recommendation model (DLRM). In addition, we described high-level architecture and efforts to enable on the conventional x86 server system. From our suggested near-memory accelerator, we got 1.94 × performance gain on two rank system which are physically multiplied.
更多
查看译文
关键词
Near-memory processing,neural network accelerator,FPGA
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要