TY - GEN
T1 - Optimizing collective operations in hybrid applications
AU - Mahéo, Aurèle
AU - Carribault, Patrick
AU - Pérache, Marc
AU - Jalby, William
N1 - Publisher Copyright:
© ACM 2014.
PY - 2014/9/9
Y1 - 2014/9/9
N2 - The advent of multicore and manycore processors in clusters advocates for combining MPI with a shared memory model like OpenMP in high-performance parallel applications. But exploiting hardware resources with such models can be sub optimal. Thus, one approach is to use the hybrid context to perform MPI communications. In this paper, we address this issue with a concept of hybrid collective communications, which consists in using OpenMP threads to parallelize MPI collectives. We validate our approach on several MPI libraries (IntelMPI and MPC), improving the overall time up to a factor of 5.29×, in a real world application.
AB - The advent of multicore and manycore processors in clusters advocates for combining MPI with a shared memory model like OpenMP in high-performance parallel applications. But exploiting hardware resources with such models can be sub optimal. Thus, one approach is to use the hybrid context to perform MPI communications. In this paper, we address this issue with a concept of hybrid collective communications, which consists in using OpenMP threads to parallelize MPI collectives. We validate our approach on several MPI libraries (IntelMPI and MPC), improving the overall time up to a factor of 5.29×, in a real world application.
KW - Collective Communications
KW - MPI
KW - OpenMP
UR - https://www.scopus.com/pages/publications/84959036891
U2 - 10.1145/2642769.2642791
DO - 10.1145/2642769.2642791
M3 - Conference contribution
AN - SCOPUS:84959036891
T3 - ACM International Conference Proceeding Series
SP - 121
EP - 122
BT - Proceedings of the 21st European MPI Users' Group Meeting, EuroMPI/ASIA 2014
PB - Association for Computing Machinery
T2 - 21st European MPI Users' Group Meeting, EuroMPI/ASIA 2014
Y2 - 9 September 2014 through 12 September 2014
ER -