DocumentCode :
244785
Title :
Massive parallelization technique for random linear network coding
Author :
Seong-Min Choi ; Joon-Sang Park
Author_Institution :
Dept. of Comput. Eng., Hongik Univ., Seoul, South Korea
fYear :
2014
fDate :
15-17 Jan. 2014
Firstpage :
296
Lastpage :
299
Abstract :
Random linear network coding (RLNC) has gain popularity as a useful performance-enhancing tool for communications networks. In this paper, we propose a RLNC parallel implementation technique for General Purpose Graphical Processing Units (GPGPUs.) Recently, GPGPU technology has paved the way for parallelizing RLNC; however, current state-of-the-art parallelization techniques for RLNC are unable to fully utilize GPGPU technology in many occasions. Addressing this problem, we propose a new RLNC parallelization technique that can fully exploit GPGPU architectures. Our parallel method shows over 4 times higher throughput compared to existing state-of-the-art parallel RLNC decoding schemes for GPGPU and 20 times higher throughput over the state-of-the-art serial RLNC decoders.
Keywords :
graphics processing units; network coding; parallel algorithms; GPGPU technology; RLNC parallelization technique; communications networks; general purpose graphical processing units; massive parallelization technique; parallel RLNC decoding schemes; performance-enhancing tool; random linear network coding; Decoding; Graphics processing units; Instruction sets; Network coding; Parallel processing; Throughput; Vectors; GPGPU; Network Coding; Parallel algorithm;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Big Data and Smart Computing (BIGCOMP), 2014 International Conference on
Conference_Location :
Bangkok
Type :
conf
DOI :
10.1109/BIGCOMP.2014.6741456
Filename :
6741456
Link To Document :
بازگشت