TY - JOUR
T1 - Breaking HPC Barriers with the 56GbE Cloud
AU - Atif, Muhammad
AU - Kobayashi, Rika
AU - Menadue, Benjamin J.
AU - Lin, Ching Yeh
AU - Sanderson, Matthew
AU - Williams, Allan
N1 - Publisher Copyright:
© 2016 The Authors. Published by Elsevier B.V.
PY - 2016
Y1 - 2016
N2 - With the widespread adoption of cloud computing, high-performance computing (HPC) is no longer limited to organisations with the funds and manpower necessary to house and run a supercomputer. However, the performance of large-scale scientific applications in the cloud has in the past been constrained by latency and bandwidth. The main reasons for these constraints are the design decisions of cloud providers, primarily focusing on high-density applications such as web services and data hosting. In this paper, we provide an overview of a high performance OpenStack cloud implementation at the National Computational Infrastructure (NCI). This cloud is targeted at high-performance scientific applications, and enables scientists to build their own clusters when their demands and software stacks conflict with traditional bare-metal HPC environments. In this paper, we present the architecture of our 56 GbE cloud and a preliminary set of HPC benchmark results against the more traditional cloud and native InfiniBand HPC environments. Three different network interconnects and configurations were tested as part of the Cloud deployment. These were 10G Ethernet, 56G Fat-tree Ethernet and native FDR Full Fat-tree InfiniBand (IB). In this paper, these three solutions are discussed from the viewpoint of on-demand HPC clusters focusing on bandwidth, latency and security. A detailed analysis of these metrics in the context of micro-benchmarks and scientific applications is presented, including the affects of using TCP and RDMA on scientific applications.
AB - With the widespread adoption of cloud computing, high-performance computing (HPC) is no longer limited to organisations with the funds and manpower necessary to house and run a supercomputer. However, the performance of large-scale scientific applications in the cloud has in the past been constrained by latency and bandwidth. The main reasons for these constraints are the design decisions of cloud providers, primarily focusing on high-density applications such as web services and data hosting. In this paper, we provide an overview of a high performance OpenStack cloud implementation at the National Computational Infrastructure (NCI). This cloud is targeted at high-performance scientific applications, and enables scientists to build their own clusters when their demands and software stacks conflict with traditional bare-metal HPC environments. In this paper, we present the architecture of our 56 GbE cloud and a preliminary set of HPC benchmark results against the more traditional cloud and native InfiniBand HPC environments. Three different network interconnects and configurations were tested as part of the Cloud deployment. These were 10G Ethernet, 56G Fat-tree Ethernet and native FDR Full Fat-tree InfiniBand (IB). In this paper, these three solutions are discussed from the viewpoint of on-demand HPC clusters focusing on bandwidth, latency and security. A detailed analysis of these metrics in the context of micro-benchmarks and scientific applications is presented, including the affects of using TCP and RDMA on scientific applications.
KW - Cloud Computing
KW - High Performance Computing
KW - High Performance Ethernet
KW - InfiniBand
KW - RDMA over Ethernet
KW - Scientific Applications
UR - http://www.scopus.com/inward/record.url?scp=84985961587&partnerID=8YFLogxK
U2 - 10.1016/j.procs.2016.07.174
DO - 10.1016/j.procs.2016.07.174
M3 - Conference article
AN - SCOPUS:84985961587
SN - 1877-0509
VL - 93
SP - 3
EP - 11
JO - Procedia Computer Science
JF - Procedia Computer Science
T2 - 6th International Conference On Advances In Computing and Communications, ICACC 2016
Y2 - 6 September 2016 through 8 September 2016
ER -