DocumentCode :
2959144
Title :
High-Performance Design of HBase with RDMA over InfiniBand
Author :
Huang, Jian ; Ouyang, Xiangyong ; Jose, Jithin ; Wasi-ur-Rahman, Md ; Wang, Hao ; Luo, Miao ; Subramoni, Hari ; Murthy, Chet ; Panda, Dhabaleswar K.
Author_Institution :
Dept. of Comput. Sci. & Eng., Ohio State Univ., Columbus, OH, USA
fYear :
2012
fDate :
21-25 May 2012
Firstpage :
774
Lastpage :
785
Abstract :
HBase is an open source distributed Key/Value store based on the idea of Big Table. It is being used in many data-center Papplications (e.g. Face book, Twitter, etc.) because of its portability and massive scalability. For this kind of system, low latency and high throughput is expected when supporting services for large scale concurrent accesses. However, the existing HBase implementation is built upon Java Sockets Interface that provides sub-optimal performance due to the overhead to provide cross-platform portability. The byte-stream oriented Java sockets semantics confine the possibility to leverage new generations of network technologies. This makes it hard to provide high performance services for data-intensive applications. High Performance Computing (HPC) domain has exploited high performance and low latency networks such as Infini Band for many years. These interconnects provide advanced network features, such as Remote Direct Memory Access (RDMA), to achieve high throughput and low latency along with low CPU utilization. RDMA follows memory-block semantics, which can be adopted efficiently to satisfy the object transmission primitives used in HBase. In this paper, we present a novel design of HBase for RDMA capable networks via Java Native Interface (JNI). Our design extends the existing open-source HBase software and makes it RDMA capable. Our performance evaluation reveals that latency of HBase Get operations of 1KB message size can be reduced to 43.7μs with the new design on QDR platform (32 Gbps). This is about a factor of 3.5 improvement over 10 Gigabit Ethernet (10 GigE) network with TCP Offload. Throughput evaluations using four HBase region servers and 64 clients indicate that the new design boosts up throughput by 3 X times over 1 GigE and 10 GigE networks. To the best of our knowledge, this is first HBase design utilizing high performance RDMA capable interconnects.
Keywords :
Java; cloud computing; distributed databases; parallel processing; programming language semantics; public domain software; software performance evaluation; software portability; user interfaces; BigTable; CPU utilization; Facebook; HBase high-performance design; HBase region clients; HBase region servers; HPC domain; InfiniBand; JNI; Java Native Interface; Java sockets interface; QDR platform; RDMA capable networks; TCP offload; Twitter; byte-stream oriented Java sockets semantics; cloud computing; cross-platform portability; data-center applications; data-intensive applications; high performance computing domain; large-scale concurrent accesses; low-latency networks; memory-block semantics; object transmission primitives; open source distributed key-value store; open-source HBase software; performance evaluation; remote direct memory access; scalability; Google; Instruction sets; Java; Servers; Sockets; Throughput; Cloud Computing and Clusters; HBase; InfiniBand; RDMA;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Parallel & Distributed Processing Symposium (IPDPS), 2012 IEEE 26th International
Conference_Location :
Shanghai
ISSN :
1530-2075
Print_ISBN :
978-1-4673-0975-2
Type :
conf
DOI :
10.1109/IPDPS.2012.74
Filename :
6267886
Link To Document :
بازگشت