EagerReuse: An Efficient Memory Reuse Approach for Complex Computational Graph.

Ruyi Qian, Bojun Cao, Mengjuan Gao, Qinwen Shi, Yida Wang,Yuanchao Xu,Qirun Huo,Keni Qiu

International Conference on Parallel and Distributed Systems(2023)

引用 0|浏览1
暂无评分
摘要
Memory reuse is a promising approach for deep neural network (DNN) to reduce memory consumption because it does not introduce any additional runtime overhead. We observe that existing memory reuse algorithms consider only the effect of an individual data feature (either tensor size or tensor lifetime) on memory reuse and ignore the relative position relationship (RPR) among tensors. As computational graphs grow slightly more complex, the mining of memory reuse becomes insufficient. To address this issue, we propose a new memory reuse algorithm—EagerReuse, which can exploit more memory reuse opportunities by analyzing RPR among tensors and reusing them as quickly as possible. We evaluated the algorithms with inference models in TensorFlow Model Garden, and the results show that the EagerReuse outperforms the state-of-the-art algorithms in three out of seven cases. For more complex computational graphs, EagerReuse can achieve better memory usage with slightly higher but acceptable overhead.
更多
查看译文
关键词
computational graph,memory optimization,memory reuse,memory usage
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要