Abstract
As major architectural changes emerge to resolve the scalability issues in many-core processors, it is critical to understand their impact on the performance of parallel programming models and run-time supports. For example, the Intel Xeon Phi KNL processor is equipped with a high-bandwidth memory and deploys a mesh-based processor interconnect. In this paper, we comprehensively analyze the impact of high-bandwidth memory and processor interconnects on the message passing interface (MPI) communication bandwidth. The results show that the bandwidth of MPI intra-node communication can be improved up to 372% by exploiting the high-bandwidth memory. In addition, we show that the bandwidth of MPI inter-node communication can be improved up to 143% with optimal core affinity. Our comprehensive study provides insight into optimization of the performance of MPI communication in emerging manycore architectures.
| Original language | English |
|---|---|
| Pages (from-to) | 170-179 |
| Number of pages | 10 |
| Journal | Journal of Computing Science and Engineering |
| Volume | 12 |
| Issue number | 4 |
| DOIs | |
| State | Published - 1 Dec 2018 |
Keywords
- Core affinity
- High-bandwidth memory
- Many-core
- MPI
- Processor interconnect
Fingerprint
Dive into the research topics of 'Exploring the performance impact of emerging many-core architectures on MPI communication'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver