A new approach to MPI collective communication implementations. Hoefler, T., Squyres, J., M., Fagg, G., E., Bosilca, G., Rehm, W., & Lumsdaine, A. Springer US, 2007.
A new approach to MPI collective communication implementations [link]Website  doi  abstract   bibtex   
Recent research into the optimization of collectiveMPIoperations has resulted in a wide variety of algorithms and corresponding implementations, each typically only applicable in a relatively narrow scope: on a specific architecture, on a specific network, with a specific number of processes, with a specific data size and/or data-type . or any combination of these (or other) factors. This situation presents an enormous challenge to portable MPI implementations which are expected to provide optimized collective operation performance on all platforms. Many portable implementations have attempted to provide a token number of algorithms that are intended to realize good performance on most systems. However, many platform configurations are still left without well-tuned collective operations. This paper presents a proposal for a framework that will allow a wide variety of collective algorithm implementations and a flexible, multi-tiered selection process for choosing which implementation to use when an application invokes an MPI collective function.
@book{
 title = {A new approach to MPI collective communication implementations},
 type = {book},
 year = {2007},
 source = {Distributed and Parallel Systems: From Cluster to Grid Computing},
 pages = {45-54},
 websites = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-50649084492&doi=10.1007%2F978-0-387-69858-8_5&partnerID=40&md5=3f3b8200e701294980d31584f677cf56},
 publisher = {Springer US},
 id = {803f348e-2988-34d6-a0de-1a74cc301048},
 created = {2018-01-09T20:30:38.831Z},
 file_attached = {false},
 profile_id = {42d295c0-0737-38d6-8b43-508cab6ea85d},
 last_modified = {2018-03-12T19:03:18.187Z},
 read = {false},
 starred = {false},
 authored = {true},
 confirmed = {true},
 hidden = {false},
 citation_key = {Hoefler200745},
 source_type = {book},
 notes = {cited By 1},
 folder_uuids = {2aba6c14-9027-4f47-8627-0902e1e2342b},
 private_publication = {false},
 abstract = {Recent research into the optimization of collectiveMPIoperations has resulted in a wide variety of algorithms and corresponding implementations, each typically only applicable in a relatively narrow scope: on a specific architecture, on a specific network, with a specific number of processes, with a specific data size and/or data-type . or any combination of these (or other) factors. This situation presents an enormous challenge to portable MPI implementations which are expected to provide optimized collective operation performance on all platforms. Many portable implementations have attempted to provide a token number of algorithms that are intended to realize good performance on most systems. However, many platform configurations are still left without well-tuned collective operations. This paper presents a proposal for a framework that will allow a wide variety of collective algorithm implementations and a flexible, multi-tiered selection process for choosing which implementation to use when an application invokes an MPI collective function.},
 bibtype = {book},
 author = {Hoefler, T and Squyres, J M and Fagg, G E and Bosilca, G and Rehm, W and Lumsdaine, A},
 doi = {10.1007/978-0-387-69858-8_5}
}

Downloads: 0