NUMA-Aware Scalable and Efficient In-Memory Aggregation on Large Domains

IEEE Trans. Knowl. Data Eng.(2015)

引用 20|浏览59
暂无评分
摘要
Business Intelligence (BI) is recognized as one of the most important IT applications in the coming big data era. In recent years, non-uniform memory access (NUMA) has become the de-facto architecture of multiprocessors on the new generation of enterprise servers. Such new architecture brings new challenges to optimization techniques on traditional operators in BI. Aggregation, for example, is one of the basic building blocks of BI, while its processing performance with existing hash-based algorithms scales poorly in terms of the number of cores under NUMA architecture. In this paper, we provide new solutions to tackle the problem of parallel hash-based aggregation, especially targeting at domains of extremely large cardinality. We propose a NUMA-aware radix partitioning (NaRP) method which divides the original huge relation table into subsets, without invoking expensive remote memory access between nodes of the cores. We also present a new efficient aggregation algorithm (EAA), to aggregate the partitioned data in parallel with low cache coherence miss and locking costs. Theoretical analysis as well as empirical study on an IBM X5 server prove that our proposals are at least two times faster than existing methods.
更多
查看译文
关键词
numa-aware radix partitioning method,optimisation,parallel processing,ibm x5 server,eaa,in-memory databases,it applications,enterprise servers,big data,cryptography,cache storage,locking costs,cache coherence miss,data partitioning,numa-aware scalable aggregation,optimization techniques,bi,narp method,competitive intelligence,numa architecture,de-facto architecture,business intelligence,hash-based algorithms,aggregation,nonuniform memory access,in-memory aggregation,memory architecture,efficient aggregation algorithm,radix-partitioning,parallel hash-based aggregation,cache miss,multiprocessors,coherence,multicore processing,servers,bismuth
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要