HOME: A Holistic GPU Memory Management Framework for Deep Learning

IEEE Transactions on Computers(2023)

引用 4|浏览22
暂无评分
摘要
We propose HOlistic MEmory management (HOME), a new framework for performing tensor placements in large DNN training when GPU memory space is not enough. HOME combines tensor swapping with tensor recomputation to reduce GPU memory footprint. Different from existing work that only considers partial DNN model information, HOME takes the holistic DNN model information into account in tensor placement decisions. More specifically, HOME uses a custom-designed particle swarm optimization algorithm to achieve the globally optimized placement for each tensor of the DNN model with a greatly reduced searching space. This holistic awareness of the whole model information enables HOME to obtain high performance under the given GPU memory constraint. We implement HOME in PyTorch and conduct our experiments using six popular DNN models. Experimental results show that HOME can outperform vDNN and Capuchin by up to 5.7x and 1.3x in throughput. Furthermore, HOME can improve the maximum batch size by up to 2.8x than the original PyTorch and up to 1.3x than Capuchin.
更多
查看译文
关键词
DNN,GPU,recomputation,swapping,tensor
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要