Abstract
The MPI is the most widespread data exchange interface standard used in parallel programming for clusters and supercomputers with many computer platforms. The primary means of the MPI communication between processes is passing messages based on basic point-to-point blocking and non-blocking routines. The choice of the optimal implementation of exchanges is essential to minimize the idle and transmission times to achieve parallel algorithm efficiency. We used three realizations of data exchange processes based on blocking, non-blocking point-to-point MPI routines and new features of the Coarray Fortran technique to determine the most efficient parallelization strategy. For the study, the two-dimensional wave equation was used as a test problem. During the experiments, the problem size and the approaches to the data exchange for transferring data between processes were changed. For each version, we measured the computation time and the acceleration factor. The research carried out shows that the larger the problem size, the greater the benefits of delayed non-blocking routines and Coarray Fortran. The efficiency of delayed non-blocking operations is due to overlapping the data transfer in the computations background. The Coarray Fortran acceleration is achieved by using Coarray variables with shared memory. The Coarray approach starts to win with the growth of problem size.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Dongarra, J., et al.: The international exascale software project roadmap. Int. J. High Perform. Comput. Appl. 25(1), 3–60 (2011)
Kostin, V.I., Lisitsa, V.V., Reshetova, G.V., Tcheverda, V.A.: Finite difference simulation of elastic waves propagation through 3D heterogeneous multiscale media based on locally refined grids. Siberian J. Num. Math. 16(1), 45–55 (2013)
Kostin, V., Lisitsa, V., Reshetova, G., Tcheverda, V.: Parallel algorithm with modulus structure for simulation of seismic wave propagation in 3D multiscale multiphysics media. In: Malyshkin, V. (ed.) PaCT 2017. LNCS, vol. 10421, pp. 42–57. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-62932-2_4
Asilah, N., Coddington, P.: Comparison of MPI benchmark programs on shared memory and distributed memory machines (point-to-point communication). Int. J. High Perform. Comput. Appl. 24(4), 469–483 (2010)
Ismail, R., Wati Abdul Hamid, N.A., Othman, M., Latip, R., Sanwani, M.A.: Point-to-point communication on gigabit ethernet and InfiniBand networks. In: Abd Manaf, A., Sahibuddin, S., Ahmad, R., Mohd Daud, S., El-Qawasmeh, E. (eds.) ICIEIS 2011. CCIS, vol. 254, pp. 369–382. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25483-3_30
Denis, A., Trahay, F.: MPI overlap: benchmark and analysis. In: 45th International Conference on Parallel Processing (ICPP), pp. 258–267. IEEE (2016). https://doi.org/10.1109/ICPP.2016.37
Kayum, N., Baddourah, A., Hajjar, O.: Methods to overlap communication with computation. In: Conference Proceedings of Third EAGE Workshop on High Performance Computing for Upstream, vol. 2017, pp. 1–10. EarthDoc (2017). https://doi.org/10.3997/2214-4609.201702326
Hoefler, T., Lumsdaine, A., Rehm, W.: Implementation and performance analysis of non-blocking collective operations for MPI. In: Proceedings of the ACM/IEEE Conference on High Performance Networking and Computing, SC 2007, pp. 1–10. Association for Computing Machinery, New York (2007). Article 52. https://doi.org/10.1145/1362622.1362692
Numrich, R.W., Reid, J.: Co-array Fortran for parallel programming. SIGPLAN Fortran Forum 17(2), 1–31 (1998)
Eachempati, D., Richardson, A., Jana, S., et al.: A Coarray Fortran implementation to support data-intensive application development. Cluster Comput. 17, 569–583 (2014). https://doi.org/10.1007/s10586-013-0302-7
Yoshimura, Ch., Bielak, J., et al.: Domain reduction method for three-dimensional earthquake modeling in localized regions, part II: verification and applications. Bull. Seismol. Soc. Am. 93, 825–840 (2003)
Moczo, P., Kristek, J., et al.: 3-D finite-difference, finite-element, discontinuous-Galerkin and spectral-element schemes analysed for their accuracy with respect to \(P\)-wave to \(S\)-wave speed ratio. Geophys. J. Int. 187, 1645–1667 (2011)
Fornberg, B.: The pseudospectral method: accurate representation of interfaces in elastic wave calculations. Geophysics 53(5), 625–637 (1988)
Takenaka, H., Wang, Y.B., et al.: An efficient approach of the pseudospectral method for modelling of geometrically symmetric seismic wavefield. Earth Planets Space 51(2), 73–79 (1999)
Chaljub, E., Komatitsch, D., et al.: Spectral-element analysis in seismology. Geophysics 50(4), 705–708 (2007)
Moczo, P., Kristek, J., et al.: 3D fourth-order staggered-grid finite-difference schemes: stability and grid dispersion. Bull. Seismol. Soc. Am. 90(3), 587–603 (2000)
Virieux, J.: \(P\)-\(SV\) wave propagation in heterogeneous media: velocity-stress finite-difference method. Geophysics 51(1), 889–901 (1986)
Levander, A.R.: Fourth-order finite-difference \(P\)-\(W\) seismograms. Geophysics 53(11), 1425–1436 (1988)
Tessmer, E.: Seismic finite-difference modeling with spatially varying time steps. Geophysics 65(4), 1290–1293 (2000)
Collino, F., Tsogka, C.: Application of the PML absorbing layer model to the linear elastodynamic problem in anisotropic heterogeneous media. Geophysics 66(1), 294–307 (2001)
TS18508: Additional Parallel Features in Fortran [Electronic resource], ISO/IEC JTC1/SC22/WG5 N 2074 (2015)
Joint Supercomputer Center of RAS. http://old.jscc.ru/eng/index.shtml
Novosibirsk Supercomputer Center of SB RAS. http://www.sscc.icmmg.nsc.ru
RS/6000 SP: Practical MPI Programming [Electronic resource]. http://wwwuser.gwdg.de/~applsw/Parallelrechner/sp_documentation/pe/sg245380.pdf
Acknowledgements
Galina Reshetova and Vladimir Cheverda have been supported by the Russian Science Foundation, project 20-11-20112.
The research is carried out using the equipment of the shared research facilities of HPC computing resources at the Joint Supercomputer Center of RAS [22] and the Siberian Supercomputer Center [23].
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Reshetova, G., Cheverda, V., Koinov, V. (2021). Comparative Efficiency Analysis of MPI Blocking and Non-blocking Communications with Coarray Fortran. In: Voevodin, V., Sobolev, S. (eds) Supercomputing. RuSCDays 2021. Communications in Computer and Information Science, vol 1510. Springer, Cham. https://doi.org/10.1007/978-3-030-92864-3_25
Download citation
DOI: https://doi.org/10.1007/978-3-030-92864-3_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-92863-6
Online ISBN: 978-3-030-92864-3
eBook Packages: Computer ScienceComputer Science (R0)