DocumentCode :
3588650
Title :
Cache-aware sparse matrix formats for Kepler GPU
Author :
Nagasaka, Yusuke ; Nukada, Akira ; Matsuoka, Satoshi
Author_Institution :
Tokyo Inst. of Technol., Tokyo, Japan
fYear :
2014
Firstpage :
281
Lastpage :
288
Abstract :
Scientific simulations often require solving extremely large sparse linear equations, whose dominant kernel is sparse matrix vector multiplication. On modern many-core processors such as GPU or MIC, the operation has been known to pose significant bottleneck and thus would result in extremely poor efficiency, because of limited processor-to-memory bandwidth and low cache hit ratio due to random access to the input vector. Our family of new sparse matrix formats for many-core processors significantly increases the cache hit ratio and thus performance by segmenting the matrix along the columns, dividing the work among the many core up to the internal cache capacity, and aggregating the result later on. Performance studies show that we achieve up to x3.0 speedup in SpMV and x1.68 in multi-node CG, compared to the best vendor libraries and competing new formats that have been recently proposed such as SELL-C-σ.
Keywords :
cache storage; graphics processing units; matrix multiplication; multiprocessing systems; Kepler GPU; cache capacity; cache hit ratio; cache-aware sparse matrix format; graphics processing unit; linear equation; many-core processor; processor-to-memory bandwidth; sparse matrix vector multiplication; Graphics processing units; Hardware; Libraries; Matrix converters; Sparse matrices; Vector processors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Parallel and Distributed Systems (ICPADS), 2014 20th IEEE International Conference on
Type :
conf
DOI :
10.1109/PADSW.2014.7097819
Filename :
7097819
Link To Document :
بازگشت