Distributed machine learning

Hyeontaek Lim, David G. Andersen,Michael Kaminsky

semanticscholar(2019)

引用 0|浏览3
暂无评分
摘要
3LC is a lossy compression scheme for state change traffic in distributed machine learning (ML) that strikes a balance between multiple goals: traffic reduction, accuracy, computation overhead, and generality. It combines three techniques—3-value quantization with sparsity multiplication, base-3 encoding, and zero-run encoding—to leverage the strengths of quantization and sparsification techniques and avoid their drawbacks. 3LC achieves a data compression ratio of up to 39–107×, preserves the high test accuracy of trained models, and provides high compression speed. Distributed ML frameworks can use 3LC without modifications to existing ML algorithms. Our experiments show that 3LC reduces wall-clock training time of ResNet-110 for CIFAR-10 on a bandwidthconstrained 10-GPU cluster by up to 16–23× compared to TensorFlow’s baseline design.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要