Change search
ReferencesLink to record
Permanent link

Direct link
Evaluation of Parallel Communication Models in Nekbone, a Nek5000 mini-application
KTH, School of Computer Science and Communication (CSC), High Performance Computing and Visualization (HPCViz).
KTH, School of Computer Science and Communication (CSC), High Performance Computing and Visualization (HPCViz). KTH, Centres, SeRC - Swedish e-Science Research Centre.ORCID iD: 0000-0002-3859-9480
KTH, School of Computer Science and Communication (CSC), High Performance Computing and Visualization (HPCViz).ORCID iD: 0000-0003-1603-5294
KTH, School of Computer Science and Communication (CSC), High Performance Computing and Visualization (HPCViz).
Show others and affiliations
2015 (English)In: 2015 IEEE International Conference on Cluster Computing, IEEE , 2015, 760-767 p.Conference paper (Refereed)
Abstract [en]

Nekbone is a proxy application of Nek5000, a scalable Computational Fluid Dynamics (CFD) code used for modelling incompressible flows. The Nekbone mini-application is used by several international co-design centers to explore new concepts in computer science and to evaluate their performance. We present the design and implementation of a new communication kernel in the Nekbone mini-application with the goal of studying the performance of different parallel communication models. First, a new MPI blocking communication kernel has been developed to solve Nekbone problems in a three-dimensional Cartesian mesh and process topology. The new MPI implementation delivers a 13% performance improvement compared to the original implementation. The new MPI communication kernel consists of approximately 500 lines of code against the original 7,000 lines of code, allowing experimentation with new approaches in Nekbone parallel communication. Second, the MPI blocking communication in the new kernel was changed to the MPI non-blocking communication. Third, we developed a new Partitioned Global Address Space (PGAS) communication kernel, based on the GPI-2 library. This approach reduces the synchronization among neighbor processes and is on average 3% faster than the new MPI-based, non-blocking, approach. In our tests on 8,192 processes, the GPI-2 communication kernel is 3% faster than the new MPI non-blocking communication kernel. In addition, we have used the OpenMP in all the versions of the new communication kernel. Finally, we highlight the future steps for using the new communication kernel in the parent application Nek5000.

Place, publisher, year, edition, pages
IEEE , 2015. 760-767 p.
National Category
Computer Science
Identifiers
URN: urn:nbn:se:kth:diva-181104DOI: 10.1109/CLUSTER.2015.131ISI: 000378648100121ScopusID: 2-s2.0-84959298440OAI: oai:DiVA.org:kth-181104DiVA: diva2:898682
Conference
WRAp2015
Note

QC 20160205

Available from: 2016-01-29 Created: 2016-01-29 Last updated: 2016-07-26Bibliographically approved

Open Access in DiVA

No full text

Other links

Publisher's full textScopus

Search in DiVA

By author/editor
Ivanov, IlyaGong, JingAkhmetova, DanaPeng, Ivy BoMarkidis, StefanoLaure, Erwin
By organisation
High Performance Computing and Visualization (HPCViz)SeRC - Swedish e-Science Research Centre
Computer Science

Search outside of DiVA

GoogleGoogle Scholar
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

Altmetric score

Total: 41 hits
ReferencesLink to record
Permanent link

Direct link