Exploiting Memory Access Patterns to Improve Memory Performance in Data-Parallel Architectures

IEEE Transactions on Parallel and Distributed Systems(2011)

引用 253|浏览3
暂无评分
摘要
The introduction of General-Purpose computation on GPUs (GPGPUs) has changed the landscape for the future of parallel computing. At the core of this phenomenon are massively multithreaded, data-parallel architectures possessing impressive acceleration ratings, offering low-cost supercomputing together with attractive power budgets. Even given the numerous benefits provided by GPGPUs, there remain a number of barriers that delay wider adoption of these architectures. One major issue is the heterogeneous and distributed nature of the memory subsystem commonly found on data-parallel architectures. Application acceleration is highly dependent on being able to utilize the memory subsystem effectively so that all execution units remain busy. In this paper, we present techniques for enhancing the memory efficiency of applications on data-parallel architectures, based on the analysis and characterization of memory access patterns in loop bodies; we target vectorization via data transformation to benefit vector-based architectures (e.g., AMD GPUs) and algorithmic memory selection for scalar-based architectures (e.g., NVIDIA GPUs). We demonstrate the effectiveness of our proposed methods with kernels from a wide range of benchmark suites. For the benchmark kernels studied, we achieve consistent and significant performance improvements (up to 11.4× and 13.5× over baseline GPU implementations on each platform, respectively) by applying our proposed methodology.
更多
查看译文
关键词
vector-based architectures,nvidia gpus,memory coalescing,computer graphic equipment,gpu computing,memory access pattern,gpu,data-parallel architecture,amd gpus,memory subsystem,data parallelism,algorithmic memory selection,parallel architectures,vectorization,memory access patterns,multi-threading,data-parallel architectures,improve memory performance,application acceleration,scalar-based architectures,memory efficiency,memory selection,exploiting memory access patterns,massive multithreaded data-parallel architectures,low-cost supercomputing,data-parallel architectures.,memory optimization,coprocessors,general-purpose computation on gpus (gpgpus),parallel computing,benchmark kernel,benchmark suite,power budgets,computer architecture,concurrent computing,acceleration,multi threading,parallel processing,kernel,logic design
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要