Small Reals Representations for Deep Learning at the Edge: A Comparison

International Conference on Next Generation Arithmetic (CoNGA)(2022)

引用 1|浏览13
暂无评分
摘要
The pervasiveness of deep neural networks (DNNs) in edge devices enforces new requirements on information representation. Low precision formats from 16 bits down to 1 or 2 bits have been proposed in the last years. In this paper we aim to illustrate a general view of the possible approaches of optimizing neural networks for DNNs at the edge. In particular we focused on these key points: i) limited non-volatile storage ii) limited volatile memory iii) limited computational power. Furthermore we explored the state-of-the-art of alternative representations for real numbers comparing their performance in recognition and detection tasks, in terms of accuracy and inference time. Finally we present our results using posits in several neural networks and datasets, showing the small accuracy degradation between 32-bit floats and 16-bit (or even 8-bit) posits, comparing the results also against the bfloat family.
更多
查看译文
关键词
Deep learning, Edge computing, Fog computing, Fine tuning at the edge, Alternative representation for real numbers, Small reals, Posits, bfloat, Weights compression
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要