Title :
Poster: An MPI Library implementing Direct Communication for Many-Core Based Accelerators
Author :
Si, Min ; Ishikawa, Yutaka
Abstract :
DCFA-MPI is an MPI library implementation for manycore-based clusters, whose compute node consists of Intel MIC (Many Integrated Core) connected to the host via PCI Express with InfiniBand. DCFA-MPI enables direct data transfer between MIC units without the host assist. MPI_Init and MPI_Finalize functions are offloaded to the host side in order to initialize the InfiniBand HCA and inform its PCI-Express address to MIC. MPI communication primitives executed in MIC may transfer data directly to other MICs or hosts by issuing commands to HCA. The implementation is based on the Mellanox InfiniBand HCA and Intel´s Knights Ferry, and compared with the Intel MPI + offload mode. Preliminary results show that DCFA-MPI outperforms the Intel MPI + offload mode by 1.12 ~ 5 times.
Keywords :
accelerator; coprocessor; direct communication; infiniband; many-core; mpi; xeon phi;
Conference_Titel :
High Performance Computing, Networking, Storage and Analysis (SCC), 2012 SC Companion:
Conference_Location :
Salt Lake City, UT
Print_ISBN :
978-1-4673-6218-4
DOI :
10.1109/SC.Companion.2012.305