Nothing Special   »   [go: up one dir, main page]

Skip to main content

Ready for the Frontier: Preparing Applications for the World’s First Exascale System

  • Conference paper
  • First Online:
High Performance Computing (ISC High Performance 2023)

Abstract

Frontier, a supercomputer at the Oak Ridge Leadership Computing Facility (OLCF), debuted atop the Top500 list of the world’s most powerful supercomputers in June 2022 as the very first computer to produce exascale performance. Making sure scientific applications are optimized on this architecture is the critical link necessary to translate the newly available computational power into scientific insight and solutions. To that goal, the OLCF developed the Center for Accelerated Application Readiness (CAAR) program to ensure that a suite of highly optimized applications is ready for scientific runs at the onset of production operations for Frontier. This paper describes our experience in porting and optimizing such suite of applications in the OLCF’s CAAR program.

Notice: This manuscript has been authored in part by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The US government retains and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript, or allow others to do so, for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    On CUDA systems, HIP acts as a thin portability layers that then calls the CUDA compiler nvcc.

  2. 2.

    How we do this can be observed from the file available from the Cholla public repository: https://github.com/cholla-hydro/cholla/blob/main/src/utils/gpu.hpp.

  3. 3.

    MI100 is a previous generation of AMD accelerators.

References

  1. AMD: New AMD ROCM information portal (2022). https://rocmdocs.amd.com/en/latest/. Accessed 01 June 2022

  2. Bussmann, M., et al.: Radiative signatures of the relativistic Kelvin-Helmholtz instability, pp. 5:1–5:12 (2013). https://doi.org/10.1145/2503210.2504564

  3. Dean, D.J., Hjorth-Jensen, M.: Coupled-cluster approach to nuclear physics. Phys. Rev. C 69, 054320 (2004). https://doi.org/10.1103/PhysRevC.69.054320

    Article  Google Scholar 

  4. Debus, A., et al.: Circumventing the dephasing and depletion limits of Laser-Wakefield acceleration. Phys. Rev. X 9, 031044 (2019). https://doi.org/10.1103/PhysRevX.9.031044

    Article  Google Scholar 

  5. Dongarra, J., Strohmaier, E., Simon, H., Meuer, M.: TOP500 (2022). https://www.top500.org/lists/top500/2022/11/

  6. Huebl, A., et al.: openPMD: a meta data standard for particle and mesh based data (2015). https://doi.org/10.5281/zenodo.591699

  7. The Open Porous Media Initiative: Open porous media project (2022). https://opm-project.org/. Accessed 30 Dec 2022

  8. Jansen, G.R.: NTCL – nuclear tensor contraction library (2022). https://gitlab.com/ntcl/ntcl

  9. Joubert, W., et al.: Accelerated application development: the ORNL titan experience. Comput. Electr. Eng. 46, 123–138 (2015). https://doi.org/10.1016/j.compeleceng.2015.04.008

    Article  Google Scholar 

  10. Joubert, W., et al.: Attacking the opioid epidemic: determining the epistatic and pleiotropic genetic architectures for chronic pain and opioid addiction. In: International Conference for High Performance Computing, Networking, Storage and Analysis, SC 2018, pp. 717–730 (2018). https://doi.org/10.1109/SC.2018.00060

  11. Kalé, L.V.: Charm++, pp. 256–264. Springer, Boston (2011)

    Google Scholar 

  12. Kohn, W., Sham, L.J.: Self-consistent equations including exchange and correlation effects. Phys. Rev. 140, A1133–A1138 (1965)

    Article  MathSciNet  Google Scholar 

  13. Lagergren, J., et al.: Climatic clustering and longitudinal analysis with impacts on food, bioenergy, and pandemics. Phytobiomes J. (2022). https://doi.org/10.1094/PBIOMES-02-22-0007-R

  14. LSMS: LSMS: scalable first principles calculations of materials using multiple scattering theory (2022). https://github.com/mstsuite/lsms. Accessed 30 Dec 2022

  15. Luo, L., et al.: Pre-exascale accelerated application development: the ORNL summit experience. IBM J. Res. Dev. 64(3/4), 11:1–11:21 (2020). https://doi.org/10.1147/JRD.2020.2965881

  16. Matthes, A., Widera, R., Zenker, E., Worpitz, B., Huebl, A., Bussmann, M.: Tuning and optimization for a variety of many-core architectures without changing a single line of implementation code using the alpaka library (2017). https://arxiv.org/abs/1706.10086

  17. McClure, J.: LBPM software package (2022). https://github.com/opm/lbpm. Accessed 30 Dec 2022

  18. McClure, J.E., Berrill, M.A., Prins, J.F., Miller, C.T.: Asynchronous in situ connected-components analysis for complex fluid flows. In: 2016 Second Workshop on In Situ Infrastructures for Enabling Extreme-Scale Analysis and Visualization (ISAV), pp. 12–17 (2016). https://doi.org/10.1109/ISAV.2016.008

  19. Oak Ridge Leadership Computing Facility: OLCF user documentation (2022). https://docs.olcf.ornl.gov/. Accessed 30 Dec 2022

  20. Korringa, J.: On the calculation of the energy of a Bloch wave in a metal. Physica 13, 392–400 (1947)

    Article  MathSciNet  Google Scholar 

  21. Perilla, J., Schulten, K.: Physical properties of the HIV-1 capsid from all-atom molecular dynamics simulations. Nat. Commun. 8(15959) (2017)

    Google Scholar 

  22. Phillips, J.C., et al.: Scalable molecular dynamics on CPU and GPU architectures with NAMD. J. Chem. Phys. 153(4), 044130 (2020)

    Article  Google Scholar 

  23. Ravikumar, K., Appelhans, D., Yeung, P.: GPU acceleration of extreme scale pseudo-spectral simulations of turbulence using asynchronism. In: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, SC 2019, pp. 1–22 (2019). https://doi.org/10.1145/3295500.3356209

  24. Schneider, E.E., Ostriker, E.C., Robertson, B.E., Thompson, T.A.: The physical nature of starburst-driven galactic outflows. Astrophys. J. 895(1), 43 (2020). https://doi.org/10.3847/1538-4357/ab8ae8

    Article  Google Scholar 

  25. Schneider, E.E., Robertson, B.E.: Cholla: a new massively parallel hydrodynamics code for astrophysical simulation. Astrophys. J. Suppl. Ser. 217(2), 24 (2015). https://doi.org/10.1088/0067-0049/217/2/24

    Article  Google Scholar 

Download references

Acknowledgments

This research used resources of the Oak Ridge Leadership Computing Facility at the Oak Ridge National Laboratory, which is supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC05-00OR22725.

The work described here were collaboratively performed by members of the respective project teams and Frontier’s Center of Excellence. We acknowledge their significant contributions to the success of the Frontier’s CAAR program. They include P.K. Yeung (Georgia Tech), Rohini Uma-Vaideswaran (Georgia Tech), Kiran Ravikumar (HPE), Steve Abbott (HPE), Matt Turner (HPE), Alessandro Fanfarillo (AMD), Swarnava Ghosh (ORNL), Kariia Karabin (ORNL), Yang Wang (Carnegie Mellon University), Vishnu Raghuraman (Carnegie Mellon University), Franco Moitzi (University of Leoben), Alessandro Fanfarillo (AMD), David Hardy (University of Illinois Urbana-Champaign), Julio Maia (AMD), Josh Vermass (Michagan State University), Tim Mattox (HPE), Morten Hjorth-Jensen (MSU), Gaute Hagen (ORNL), Justin Lietz (ORNL), Rene Widera (Helmholtz-Zentrum Dresden-Rossendorf - HZDR), Klaus Steiniger (HZDR), Sergei Bastrakov (HZDR), Michael Bussmann (HZDR), Fabian Mora (U. Delaware), Richard Pausch (HZDR), Guido Juckeland (HZDR), Jeffrey Kelling (HZDR), Matthew Leinhauser (U. Delaware), Jeffery Young (Georgia Tech.), Franz Pöschl (HZDR), Alexander Debus (HZDR), Sunita Chandrasekaran (U. Delaware), Evan Schneider (U. Pittsburgh), Bruno Villasenor (U. California Santa Cruz, AMD), Brant Robertson ((U. California Santa Cruz), Robert Caddy (U. Pittsburgh), Alwin Mao (U. Pittsburgh), Trey White (HPE), Dan Jacobson (ORNL), Jakub Kurzak (AMD).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Reuben D. Budiardja .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Budiardja, R.D. et al. (2023). Ready for the Frontier: Preparing Applications for the World’s First Exascale System. In: Bhatele, A., Hammond, J., Baboulin, M., Kruse, C. (eds) High Performance Computing. ISC High Performance 2023. Lecture Notes in Computer Science, vol 13948. Springer, Cham. https://doi.org/10.1007/978-3-031-32041-5_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-32041-5_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-32040-8

  • Online ISBN: 978-3-031-32041-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics