Abstract
As major architectural changes emerge to resolve the scalability issues in many-core processors, it is critical to understand their impact on the performance of parallel programming models and run-time supports. For example, the Intel Xeon Phi KNL processor is equipped with a high-bandwidth memory and deploys a mesh-based processor interconnect. In this paper, we comprehensively analyze the impact of high-bandwidth memory and processor interconnects on the message passing interface (MPI) communication bandwidth. The results show that the bandwidth of MPI intra-node communication can be improved up to 372% by exploiting the high-bandwidth memory. In addition, we show that the bandwidth of MPI inter-node communication can be improved up to 143% with optimal core affinity. Our comprehensive study provides insight into optimization of the performance of MPI communication in emerging manycore architectures.
Original language | English |
---|---|
Pages (from-to) | 170-179 |
Number of pages | 10 |
Journal | Journal of Computing Science and Engineering |
Volume | 12 |
Issue number | 4 |
DOIs | |
State | Published - 1 Dec 2018 |
Keywords
- Core affinity
- High-bandwidth memory
- Many-core
- MPI
- Processor interconnect