کد مقاله کد نشریه سال انتشار مقاله انگلیسی نسخه تمام متن
432655 689006 2016 21 صفحه PDF دانلود رایگان
عنوان انگلیسی مقاله ISI
Optimization techniques for sparse matrix–vector multiplication on GPUs
موضوعات مرتبط
مهندسی و علوم پایه مهندسی کامپیوتر نظریه محاسباتی و ریاضیات
پیش نمایش صفحه اول مقاله
Optimization techniques for sparse matrix–vector multiplication on GPUs
چکیده انگلیسی


• We propose AdELL+, an advanced GPU-based sparse matrix format.
• We integrate complementary optimizations such as adaptivity and compression.
• We include a warp-balancing heuristic and auto-tuning approach into a cohesive framework.
• Our analysis shows that AdELL+ consistently outperforms the previous related works.

Sparse linear algebra is fundamental to numerous areas of applied mathematics, science and engineering. In this paper, we propose an efficient data structure named AdELL+ for optimizing the SpMV kernel on GPUs, focusing on performance bottlenecks of sparse computation. The foundation of our work is an ELL-based adaptive format which copes with matrix irregularity using balanced warps composed using a parametrized warp-balancing heuristic. We also address the intrinsic bandwidth-limited nature of SpMV with warp granularity, blocking, delta compression and nonzero unrolling, targeting both memory footprint and memory hierarchy efficiency. Finally, we introduce a novel online auto-tuning approach that uses a quality metric to predict efficient block factors and that hides preprocessing overhead with useful SpMV computation. Our experimental results show that AdELL+ achieves comparable or better performance over other state-of-the-art SpMV sparse formats proposed in academia (BCCOO) and industry (CSR+ and CSR-Adaptive). Moreover, our auto-tuning approach makes AdELL+ viable for real-world applications.

ناشر
Database: Elsevier - ScienceDirect (ساینس دایرکت)
Journal: Journal of Parallel and Distributed Computing - Volumes 93–94, July 2016, Pages 66–86
نویسندگان
, ,