A distributed kernel summation framework for general-dimension machine learning

Periodicals(2014)

引用 36|浏览318
暂无评分
摘要
AbstractKernel summations are a ubiquitous key computational bottleneck in many data analysis methods. In this paper, we attempt to marry, for the first time, the best relevant techniques in parallel computing, where kernel summations are in low dimensions, with the best general-dimension algorithms from the machine learning literature. We provide the first distributed implementation of kernel summation framework that can utilize: i various types of deterministic and probabilistic approximations that may be suitable for low and high-dimensional problems with a large number of data points; ii any multidimensional binary tree using both distributed memory and shared memory parallelism; and iii a dynamic load balancing scheme to adjust work imbalances during the computation. Our hybrid message passing interface MPI/OpenMP codebase has wide applicability in providing a general framework to accelerate the computation of many popular machine learning methods. Our experiments show scalability results for kernel density estimation on a synthetic ten-dimensional dataset containing over one billion points and a subset of the Sloan Digital Sky Survey Data up to 6144 cores. © 2013 Wiley Periodicals, Inc. Statistical Analysis and Data Mining, 2013
更多
查看译文
关键词
kernel methods,nonparametric methods,parallel machine learning,GPGPU,parallel multidimensional trees,CUDA
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要