Abstract
Shared-memory and message-passing are two opposite models to develop parallel computations. The shared-memory model, adopted by existing frameworks such as OpenMP, represents a de-facto standard on multi-/many-core architectures. However, message-passing deserves to be studied for its inherent properties in terms of portability and flexibility as well as for its better ease of debugging. Achieving good performance from the use of messages in shared-memory architectures requires an efficient implementation of the run-time support. This paper investigates the definition of a delegation mechanism on multi-threaded architectures able to: (i) overlap communications with calculation phases, (ii) parallelize distribution and collective operations. Our ideas have been exemplified using two parallel benchmarks on the Intel Phi, showing that in these applications our message-passing support outperforms MPI and reaches similar performance compared to standard OpenMP implementations. © 2014 IEEE.
Original language | English |
---|---|
Title of host publication | Proceedings - 2014 22nd Euromicro International Conference on Parallel, Distributed, and Network-Based Processing, PDP 2014 |
Publisher | IEEE Computer Society |
Pages | 262-270 |
ISBN (Print) | 9781479927289 |
DOIs | |
Publication status | Published - 1 Jan 2014 |
Externally published | Yes |
Event | 2014 22nd Euromicro International Conference on Parallel, Distributed, and Network-Based Processing, PDP 2014 - , Italy Duration: 12 Feb 2014 → 14 Feb 2014 |
Conference
Conference | 2014 22nd Euromicro International Conference on Parallel, Distributed, and Network-Based Processing, PDP 2014 |
---|---|
Country/Territory | Italy |
Period | 12/02/14 → 14/02/14 |