{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,3]],"date-time":"2025-03-03T05:34:50Z","timestamp":1740980090676,"version":"3.38.0"},"reference-count":18,"publisher":"SAGE Publications","issue":"3","license":[{"start":{"date-parts":[[2018,11,14]],"date-time":"2018-11-14T00:00:00Z","timestamp":1542153600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"funder":[{"DOI":"10.13039\/100010664","name":"H2020 Future and Emerging Technologies","doi-asserted-by":"publisher","award":["INTERTWinE project (no. 671602)"],"award-info":[{"award-number":["INTERTWinE project (no. 671602)"]}],"id":[{"id":"10.13039\/100010664","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["The International Journal of High Performance Computing Applications"],"published-print":{"date-parts":[[2019,5]]},"abstract":"<jats:p> One of the main hurdles of partitioned global address space (PGAS) approaches is the dominance of message passing interface (MPI), which as a de facto standard appears in the code basis of many applications. To take advantage of the PGAS APIs like global address space programming interface (GASPI) without a major change in the code basis, interoperability between MPI and PGAS approaches needs to be ensured. In this article, we consider an interoperable GASPI\/MPI implementation for the communication\/performance crucial parts of the Ludwig and iPIC3D applications. To address the discovered performance limitations, we develop a novel strategy for significantly improved performance and interoperability between both APIs by leveraging GASPI shared windows and shared notifications. First results with a corresponding implementation in the MiniGhost proxy application and the Allreduce collective operation demonstrate the viability of this approach. <\/jats:p>","DOI":"10.1177\/1094342018808359","type":"journal-article","created":{"date-parts":[[2018,11,15]],"date-time":"2018-11-15T03:56:48Z","timestamp":1542254208000},"page":"554-568","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":1,"title":["Interoperability strategies for GASPI and MPI in large-scale scientific applications"],"prefix":"10.1177","volume":"33","author":[{"given":"Christian","family":"Simmendinger","sequence":"first","affiliation":[{"name":"T-Systems Solutions for Research, Stuttgart, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2414-700X","authenticated-orcid":false,"given":"Roman","family":"Iakymchuk","sequence":"additional","affiliation":[{"name":"KTH Royal Institute of Technology, Stockholm, Sweden"}]},{"given":"Luis","family":"Cebamanos","sequence":"additional","affiliation":[{"name":"EPCC, The University of Edinburgh, Edinburgh, UK"}]},{"given":"Dana","family":"Akhmetova","sequence":"additional","affiliation":[{"name":"KTH Royal Institute of Technology, Stockholm, Sweden"}]},{"given":"Valeria","family":"Bartsch","sequence":"additional","affiliation":[{"name":"Fraunhofer ITWM, Kaiserslautern, Germany"}]},{"given":"Tiberiu","family":"Rotaru","sequence":"additional","affiliation":[{"name":"Fraunhofer ITWM, Kaiserslautern, Germany"}]},{"given":"Mirko","family":"Rahn","sequence":"additional","affiliation":[{"name":"Fraunhofer ITWM, Kaiserslautern, Germany"}]},{"given":"Erwin","family":"Laure","sequence":"additional","affiliation":[{"name":"KTH Royal Institute of Technology, Stockholm, Sweden"}]},{"given":"Stefano","family":"Markidis","sequence":"additional","affiliation":[{"name":"KTH Royal Institute of Technology, Stockholm, Sweden"}]}],"member":"179","published-online":{"date-parts":[[2018,11,14]]},"reference":[{"key":"bibr1-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-78054-2_26"},{"volume-title":"Minighost: A Miniapp for Exploring Boundary Exchange Strategies using Stencil Computations in Scientific Parallel Computing","year":"2011","author":"Barrett RF","key":"bibr2-1094342018808359"},{"key":"bibr3-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1109\/IPDPS.2015.30"},{"key":"bibr4-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1145\/2802658.2802669"},{"issue":"3","key":"bibr5-1094342018808359","doi-asserted-by":"crossref","first-page":"3737","DOI":"10.1029\/1999JA001001","volume":"106","author":"Birn J","year":"2001","journal-title":"JGR: Space Physics"},{"key":"bibr6-1094342018808359","unstructured":"Davidson E (2008) Message-Passing for Lattice Boltzmann. Master\u2019s Thesis, EPCC, The University of Edinburgh, Scotland, UK."},{"key":"bibr7-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1016\/S0010-4655(00)00205-8"},{"issue":"3","key":"bibr8-1094342018808359","first-page":"274","volume":"29","author":"Gray A","year":"2015","journal-title":"IJHPCA"},{"volume-title":"Guide to porting MPI applications to GPI-2. Technical report","year":"2015","author":"Machado R","key":"bibr9-1094342018808359"},{"key":"bibr10-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1016\/j.jcp.2014.02.002"},{"key":"bibr11-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1016\/j.matcom.2009.08.038"},{"key":"bibr12-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46079-6_5"},{"key":"bibr13-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1016\/0734-743X(90)90071-3"},{"key":"bibr14-1094342018808359","unstructured":"MVAPICH (2018) MVAPICH: MPI over InfiniBand, omni-path, ethernet\/iWARP, and RoCE. http:\/\/mvapich.cse.ohio-state.edu\/benchmarks\/ (accessed 2 December 2018)."},{"key":"bibr15-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1145\/3127024.3127034"},{"key":"bibr16-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1137\/1.9780898718003"},{"key":"bibr17-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1109\/HOTI.2015.13"},{"key":"bibr18-1094342018808359","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10626-7_2"}],"container-title":["The International Journal of High Performance Computing Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.1177\/1094342018808359","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.1177\/1094342018808359","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.1177\/1094342018808359","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,2]],"date-time":"2025-03-02T10:05:06Z","timestamp":1740909906000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.1177\/1094342018808359"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,11,14]]},"references-count":18,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2019,5]]}},"alternative-id":["10.1177\/1094342018808359"],"URL":"https:\/\/doi.org\/10.1177\/1094342018808359","relation":{},"ISSN":["1094-3420","1741-2846"],"issn-type":[{"type":"print","value":"1094-3420"},{"type":"electronic","value":"1741-2846"}],"subject":[],"published":{"date-parts":[[2018,11,14]]}}}