Parallel deep neural network training for big data on blue gene/Q

IEEE Trans. Parallel Distrib. Syst.(2017)

引用 66|浏览265
暂无评分
摘要
Deep Neural Networks (DNNs) have recently been shown to significantly outperform existing machine learning techniques in several pattern recognition tasks. DNNs are the state-of-the-art models used in image recognition, object detection, classification and tracking, and speech and language processing applications. The biggest drawback to DNNs has been the enormous cost in computation and time taken to train the parameters of the networks - often a tenfold increase relative to conventional technologies. Such training time costs can be mitigated by the application of parallel computing algorithms and architectures. However, these algorithms often run into difficulties because of the cost of inter-processor communication bottlenecks. In this paper, we describe how to enable Parallel Deep Neural Network Training on the IBM Blue Gene/Q (BG/Q) computer system. Specifically, we explore DNN training using the data-parallel Hessian-free 2nd order optimization algorithm. Such an algorithm is particularly well-suited to parallelization across a large set of loosely coupled processors. BG/Q, with its excellent inter-processor communication characteristics, is an ideal match for this type of algorithm. The paper discusses how issues regarding programming model and data-dependent imbalances are addressed. Results on large-scale speech tasks show that the performance on BG/Q scales linearly up to 4096 processes with no loss in accuracy. This allows us to train neural networks using billions of training examples in a few hours.
更多
查看译文
关键词
dnn,ibm bg/q computer system,ibm blue gene/q computer system,pattern recognition,big data,pattern recognition tasks,speech recognition,parallel computing algorithms,learning (artificial intelligence),data-parallel hessian-free 2nd order optimization algorithm,parallel architectures,machine learning techniques,high performance computing,training time costs,parallel deep neural network training,programming model,interprocessor communication characteristics,neural nets
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要