Evaluation of MPI Allreduce for Distributed Training of Convolutional Neural Networks

2021 29th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP)(2021)

引用 2|浏览15
暂无评分
摘要
Training deep neural networks is a costly procedure, often performed via sophisticated deep learning frameworks on clusters of computers. As faster processor technologies are integrated into these cluster facilities (e.g., NVIDIA’s graphics accelerators or Google’s tensor processing units), the communication component of the training process rapidly becomes a performance bottleneck. In this paper,...
更多
查看译文
关键词
Training,Graphics,Deep learning,Tensors,Message passing,Neural networks,Convolutional neural networks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要