Abstract
The advent of multi-/many-core architectures demands efficient run-time supports to sustain parallel applications scalability. Synchronization mechanisms should be optimized in order to account for different scenarios, such as the interaction between threads executed on different cores as well as intra-core synchronization, i.e. involving threads executed on hardware contexts of the same core. In this perspective, we describe the design issues of two notable mechanisms for shared-memory parallel computations. We point out how specific architectural supports, like hardware cache coherence and core-to-core interconnection networks, make it possible to design optimized implementations of such mechanisms. In this paper we discuss experimental results on three representative architectures: a flagship Intel multi-core and two interesting network processors. The final result helps to untangle the complex implementation space of synchronization mechanisms.
Original language | English |
---|---|
Title of host publication | IASTED Multiconferences - Proceedings of the IASTED International Conference on Parallel and Distributed Computing and Networks, PDCN 2013 |
Pages | 576-585 |
DOIs | |
Publication status | Published - 2013 |
Externally published | Yes |
Event | 11th IASTED International Conference on Parallel and Distributed Computing and Networks, PDCN 2013 - , Austria Duration: 11 Feb 2013 → 13 Feb 2013 |
Conference
Conference | 11th IASTED International Conference on Parallel and Distributed Computing and Networks, PDCN 2013 |
---|---|
Country/Territory | Austria |
Period | 11/02/13 → 13/02/13 |