Abstract
Frontier, a supercomputer at the Oak Ridge Leadership Computing Facility (OLCF), debuted atop the Top500 list of the world’s most powerful supercomputers in June 2022 as the very first computer to produce exascale performance. Making sure scientific applications are optimized on this architecture is the critical link necessary to translate the newly available computational power into scientific insight and solutions. To that goal, the OLCF developed the Center for Accelerated Application Readiness (CAAR) program to ensure that a suite of highly optimized applications is ready for scientific runs at the onset of production operations for Frontier. This paper describes our experience in porting and optimizing such suite of applications in the OLCF’s CAAR program.
Notice: This manuscript has been authored in part by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The US government retains and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript, or allow others to do so, for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
On CUDA systems, HIP acts as a thin portability layers that then calls the CUDA compiler nvcc.
- 2.
How we do this can be observed from the file available from the Cholla public repository: https://github.com/cholla-hydro/cholla/blob/main/src/utils/gpu.hpp.
- 3.
MI100 is a previous generation of AMD accelerators.
References
AMD: New AMD ROCM information portal (2022). https://rocmdocs.amd.com/en/latest/. Accessed 01 June 2022
Bussmann, M., et al.: Radiative signatures of the relativistic Kelvin-Helmholtz instability, pp. 5:1–5:12 (2013). https://doi.org/10.1145/2503210.2504564
Dean, D.J., Hjorth-Jensen, M.: Coupled-cluster approach to nuclear physics. Phys. Rev. C 69, 054320 (2004). https://doi.org/10.1103/PhysRevC.69.054320
Debus, A., et al.: Circumventing the dephasing and depletion limits of Laser-Wakefield acceleration. Phys. Rev. X 9, 031044 (2019). https://doi.org/10.1103/PhysRevX.9.031044
Dongarra, J., Strohmaier, E., Simon, H., Meuer, M.: TOP500 (2022). https://www.top500.org/lists/top500/2022/11/
Huebl, A., et al.: openPMD: a meta data standard for particle and mesh based data (2015). https://doi.org/10.5281/zenodo.591699
The Open Porous Media Initiative: Open porous media project (2022). https://opm-project.org/. Accessed 30 Dec 2022
Jansen, G.R.: NTCL – nuclear tensor contraction library (2022). https://gitlab.com/ntcl/ntcl
Joubert, W., et al.: Accelerated application development: the ORNL titan experience. Comput. Electr. Eng. 46, 123–138 (2015). https://doi.org/10.1016/j.compeleceng.2015.04.008
Joubert, W., et al.: Attacking the opioid epidemic: determining the epistatic and pleiotropic genetic architectures for chronic pain and opioid addiction. In: International Conference for High Performance Computing, Networking, Storage and Analysis, SC 2018, pp. 717–730 (2018). https://doi.org/10.1109/SC.2018.00060
Kalé, L.V.: Charm++, pp. 256–264. Springer, Boston (2011)
Kohn, W., Sham, L.J.: Self-consistent equations including exchange and correlation effects. Phys. Rev. 140, A1133–A1138 (1965)
Lagergren, J., et al.: Climatic clustering and longitudinal analysis with impacts on food, bioenergy, and pandemics. Phytobiomes J. (2022). https://doi.org/10.1094/PBIOMES-02-22-0007-R
LSMS: LSMS: scalable first principles calculations of materials using multiple scattering theory (2022). https://github.com/mstsuite/lsms. Accessed 30 Dec 2022
Luo, L., et al.: Pre-exascale accelerated application development: the ORNL summit experience. IBM J. Res. Dev. 64(3/4), 11:1–11:21 (2020). https://doi.org/10.1147/JRD.2020.2965881
Matthes, A., Widera, R., Zenker, E., Worpitz, B., Huebl, A., Bussmann, M.: Tuning and optimization for a variety of many-core architectures without changing a single line of implementation code using the alpaka library (2017). https://arxiv.org/abs/1706.10086
McClure, J.: LBPM software package (2022). https://github.com/opm/lbpm. Accessed 30 Dec 2022
McClure, J.E., Berrill, M.A., Prins, J.F., Miller, C.T.: Asynchronous in situ connected-components analysis for complex fluid flows. In: 2016 Second Workshop on In Situ Infrastructures for Enabling Extreme-Scale Analysis and Visualization (ISAV), pp. 12–17 (2016). https://doi.org/10.1109/ISAV.2016.008
Oak Ridge Leadership Computing Facility: OLCF user documentation (2022). https://docs.olcf.ornl.gov/. Accessed 30 Dec 2022
Korringa, J.: On the calculation of the energy of a Bloch wave in a metal. Physica 13, 392–400 (1947)
Perilla, J., Schulten, K.: Physical properties of the HIV-1 capsid from all-atom molecular dynamics simulations. Nat. Commun. 8(15959) (2017)
Phillips, J.C., et al.: Scalable molecular dynamics on CPU and GPU architectures with NAMD. J. Chem. Phys. 153(4), 044130 (2020)
Ravikumar, K., Appelhans, D., Yeung, P.: GPU acceleration of extreme scale pseudo-spectral simulations of turbulence using asynchronism. In: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, SC 2019, pp. 1–22 (2019). https://doi.org/10.1145/3295500.3356209
Schneider, E.E., Ostriker, E.C., Robertson, B.E., Thompson, T.A.: The physical nature of starburst-driven galactic outflows. Astrophys. J. 895(1), 43 (2020). https://doi.org/10.3847/1538-4357/ab8ae8
Schneider, E.E., Robertson, B.E.: Cholla: a new massively parallel hydrodynamics code for astrophysical simulation. Astrophys. J. Suppl. Ser. 217(2), 24 (2015). https://doi.org/10.1088/0067-0049/217/2/24
Acknowledgments
This research used resources of the Oak Ridge Leadership Computing Facility at the Oak Ridge National Laboratory, which is supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC05-00OR22725.
The work described here were collaboratively performed by members of the respective project teams and Frontier’s Center of Excellence. We acknowledge their significant contributions to the success of the Frontier’s CAAR program. They include P.K. Yeung (Georgia Tech), Rohini Uma-Vaideswaran (Georgia Tech), Kiran Ravikumar (HPE), Steve Abbott (HPE), Matt Turner (HPE), Alessandro Fanfarillo (AMD), Swarnava Ghosh (ORNL), Kariia Karabin (ORNL), Yang Wang (Carnegie Mellon University), Vishnu Raghuraman (Carnegie Mellon University), Franco Moitzi (University of Leoben), Alessandro Fanfarillo (AMD), David Hardy (University of Illinois Urbana-Champaign), Julio Maia (AMD), Josh Vermass (Michagan State University), Tim Mattox (HPE), Morten Hjorth-Jensen (MSU), Gaute Hagen (ORNL), Justin Lietz (ORNL), Rene Widera (Helmholtz-Zentrum Dresden-Rossendorf - HZDR), Klaus Steiniger (HZDR), Sergei Bastrakov (HZDR), Michael Bussmann (HZDR), Fabian Mora (U. Delaware), Richard Pausch (HZDR), Guido Juckeland (HZDR), Jeffrey Kelling (HZDR), Matthew Leinhauser (U. Delaware), Jeffery Young (Georgia Tech.), Franz Pöschl (HZDR), Alexander Debus (HZDR), Sunita Chandrasekaran (U. Delaware), Evan Schneider (U. Pittsburgh), Bruno Villasenor (U. California Santa Cruz, AMD), Brant Robertson ((U. California Santa Cruz), Robert Caddy (U. Pittsburgh), Alwin Mao (U. Pittsburgh), Trey White (HPE), Dan Jacobson (ORNL), Jakub Kurzak (AMD).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Budiardja, R.D. et al. (2023). Ready for the Frontier: Preparing Applications for the World’s First Exascale System. In: Bhatele, A., Hammond, J., Baboulin, M., Kruse, C. (eds) High Performance Computing. ISC High Performance 2023. Lecture Notes in Computer Science, vol 13948. Springer, Cham. https://doi.org/10.1007/978-3-031-32041-5_10
Download citation
DOI: https://doi.org/10.1007/978-3-031-32041-5_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-32040-8
Online ISBN: 978-3-031-32041-5
eBook Packages: Computer ScienceComputer Science (R0)