SpV8: Pursuing Optimal Vectorization and Regular Computation Pattern in SpMV

2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC)(2021)

引用 6|浏览26
暂无评分
摘要
Sparse Matrix-Vector Multiplication (SpMV) plays an important role in many scientific and industry applications, and remains a well-known challenge due to the high sparsity and irregularity. Most existing researches on SpMV try to pursue high vectorization efficiency. However, such approaches may suffer from non-negligible speculation penalty due to their irregular computation patterns. In this paper, we propose SpV8, a novel approach that optimizes both speculation and vectorization in SpMV. Specifically, SpV8 analyzes data distribution in different matrices and row panels, and accordingly applies optimization method that achieves the maximal vectorization with regular computation patterns. We evaluate SpV8 on Intel Xeon CPU and compare with multiple state-of-art SpMV algorithms using 71 sparse matrices. The results show that SpV8 achieves up to 10 x speedup (average 2.8 x) against the standard MKL SpMV routine, and up to 2.4 x speedup (average 1.4 x) against the best existing approach. Moreover, SpMV features very low preprocessing overhead in all compared approaches, which indicates SpV8 is highly-applicable in real-world applications.
更多
查看译文
关键词
SpMV, SIMD, Vectorization, Speculation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要