Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3240765.3243484guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
research-article

Machine Learning for Performance and Power Modeling of Heterogeneous Systems

Published: 05 November 2018 Publication History

Abstract

Modern processing systems with heterogeneous components (e.g., CPUs, GPUs) have numerous configuration and design options such as the number and types of cores, frequency, and memory bandwidth. Hardware architects must perform design space explorations in order to accurately target markets of interest under tight time-to-market constraints. This need highlights the importance of rapid performance and power estimation mechanisms. This work describes the use of machine learning (ML) techniques within a methodology for the estimating performance and power of heterogeneous systems. In particular, we measure the power and performance of a large collection of test applications running on real hardware across numerous hardware configurations. We use these measurements to train a ML model; the model learns how the applications scale with the system's key design parameters. Later, new applications of interest are executed on a single configuration, and we gather hardware performance counter values which describe how the application used the hardware. These values are fed into our ML model's inference algorithm, which quickly identify how this application will scale across various design points. In this way, we can rapidly predict the performance and power of the new application across a wide range of system configurations. Once the initial run of the program is complete, our ML algorithm can predict the application's performance and power at many hardware points faster than running it at each of those points and with a level of accuracy comparable to cycle-level simulators.

References

[1]
Advanced Micro Devices, Inc. 2018. AMD A-Series APU Processors. https://www.amd.com/en/products/a-series-processors-desktop-7th-gen-fm2-plus
[2]
Advanced Micro Devices, Inc. 2018. AMD Radeon™ R9 Series Graphics. http://www.amd.com/en-us/products/graphics/desktop/r9
[3]
Sebastian Anthony. 2013. Xbox One vs. PS4: How the Final Hardware Specs Compare. http://www.extremetech.com/gaming/156273-xbox-720-vs-ps4-vs-pc-how-the-hardware-specs-compare
[4]
David Brooks, Mark Hempstead, Mike Lui, Parnian Mokri, Siddharth Nilakantan, Brandon Reagen, and Yakun Sophia Shao. 2015. Research Infrastructures for Accelerator-Centric Architectures. Tutorial Presented at HPCA.
[5]
Shuai Che, J.W. Sheaffer, M. Boyer, L.G. Szafaryn, Liang Wang, and K. Skadron. 2010. A Characterization of the Rodinia Benchmark Suite with Comparison to Contemporary CMP Workloads. In Proc. of the IEEE Int'l Symp. on Workload Characterization (IISWC).
[6]
Joseph L. Greathouse and Mayank Daga. 2014. Efficient Sparse Matrix-Vector Multiplication on GPUs using the CSR Storage Format. In Proc. of the Conference on High Performance Computing, Networking, Storage and Analysis (SC).
[7]
Joseph L. Greathouse, Alexander Lyashevsky, Mitesh Meswani, Nuwan Jayasena, and Michael Ignatowski. 2013. Simulation of Exascale Nodes through Runtime Hardware Monitoring. In ASCR Workshop on Modeling & Simulation of Exascale Systems & Applications (ModSim).
[8]
Anthony Gutierrez, Joseph Pusdesris, Ronald G. Dreslinski, Trevor Mudge, Chander Sudanthi, Christopher D. Emmons, Mitchell Hayenga, and Nigel Paver. 2014. Sources of Error in Full-System Simulation. In Proc. of the Int'l Symp. on Performance Analysis of Systems and Software (ISPASS).
[10]
Wenhao Jia, Kelly A. Shaw, and Margaret Martonosi. 2013. Stargazer: Automated Regression-Based GPU Design Space Exploration. In Proc. of the Int'l Symp. on Performance Analysis of Systems and Software (ISPASS).
[11]
Abhinandan Majumdar, Gene Wu, Kapil Dev, Joseph L. Greathouse, Indrani Paul, Wei Huang, Arjun Karthik Venugopal, Leonardo Piga, Chip Freitag, and Sooraj Puthoor. 2015. A Taxonomy of GPGPU Performance Scaling. In Proc. of the IEEE Int'l Symp. on Workload Characterization (IISWC).
[12]
Aashish Phansalkar, Ajay Joshi, and Lizy K. John. 2007. Subsetting the SPEC CPU2006 Benchmark Suite. SIGARCH Computer Architecture News 35, 1 (2007), 69–76.
[13]
Timothy Sherwood, Erez Perelman, Greg Hamerly, and Brad Calder. 2002. Automatically Characterizing Large Scale Program Behavior. In Proc. of the Int'l Conf. on Architectural Support for Programming Languages and Operating Systems (ASPLOS).
[14]
C. Spearman. 1907. Demonstration of Formulæfor True Measurement of Correlation. The American Journal of Psychology 18, 2 (1907), 161–169.
[15]
Bo Su, Joseph L. Greathouse, Junli Gu, Michael Boyer, Li Shen, and Zhiying Wang. 2014. Implementing a Leading Loads Performance Predictor on Commodity Procesors. In Proc. of the USENIX Annual Technical Conf. (USENIX ATC).
[16]
Bo Su, Junli Gu, Li Shen, Wei Huang, Joseph L. Greathouse, and Zhiying Wang. 2014. PPEP: Online Performance, Power, and Energy Prediction Framework and DVFS Space Exploration. In Proc. of the Int'l Symp. on Microarchitecture (MICRO).
[17]
Joe H. Ward. 1963. Hierarchical Grouping to Optimize an Objective Function. J. Amer. Statist. Assoc. 58, 301 (1963), 236–244.
[18]
Samuel Williams, Andrew Watterman, and David Patterson. 2009. Roofline: an Insightful Visual Performance Model for Multicore Architectures. Commun. ACM 52, 4 (April 2009), 56–76.
[19]
Gene Wu, Toseph L. Greathouse, Alexander Lyashevsky, Nuwan Jayasena, and Derek Chiou. 2015. GPGPU Performance and Power Estimation Using Machine Learning. In Proc. of the Int'l Symp. on High Performance Computer Architecture (HPCA).

Cited By

View all
  • (2023)HPC Application Performance Prediction with Machine Learning on New ArchitecturesProceedings of the 2023 on Performance EngineeRing, Modelling, Analysis, and VisualizatiOn Strategy10.1145/3588993.3597262(1-8)Online publication date: 28-Jul-2023
  • (2022)Predicting physical computer systems performance and power from simulation systems using machine learning modelComputing10.1007/s00607-022-01066-5105:5(935-953)Online publication date: 15-Mar-2022
  • (2022)Orchestrated Co-scheduling, Resource Partitioning, and Power Capping on CPU-GPU Heterogeneous Systems via Machine LearningArchitecture of Computing Systems10.1007/978-3-031-21867-5_4(51-67)Online publication date: 14-Dec-2022
  • Show More Cited By

Index Terms

  1. Machine Learning for Performance and Power Modeling of Heterogeneous Systems
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Please enable JavaScript to view thecomments powered by Disqus.

          Information & Contributors

          Information

          Published In

          cover image Guide Proceedings
          2018 IEEE/ACM International Conference on Computer-Aided Design (ICCAD)
          Nov 2018
          939 pages

          Publisher

          IEEE Press

          Publication History

          Published: 05 November 2018

          Permissions

          Request permissions for this article.

          Qualifiers

          • Research-article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 10 Nov 2024

          Other Metrics

          Citations

          Cited By

          View all
          • (2023)HPC Application Performance Prediction with Machine Learning on New ArchitecturesProceedings of the 2023 on Performance EngineeRing, Modelling, Analysis, and VisualizatiOn Strategy10.1145/3588993.3597262(1-8)Online publication date: 28-Jul-2023
          • (2022)Predicting physical computer systems performance and power from simulation systems using machine learning modelComputing10.1007/s00607-022-01066-5105:5(935-953)Online publication date: 15-Mar-2022
          • (2022)Orchestrated Co-scheduling, Resource Partitioning, and Power Capping on CPU-GPU Heterogeneous Systems via Machine LearningArchitecture of Computing Systems10.1007/978-3-031-21867-5_4(51-67)Online publication date: 14-Dec-2022
          • (2021)Real-Time Full-Chip Thermal Tracking: A Post-Silicon, Machine Learning PerspectiveIEEE Transactions on Computers10.1109/TC.2021.3086112(1-1)Online publication date: 2021
          • (2021)Can We Trust Machine Learning for Electronic Design Automation?2021 IEEE 34th International System-on-Chip Conference (SOCC)10.1109/SOCC52499.2021.9739485(135-140)Online publication date: 14-Sep-2021
          • (2021)Fast and Accurate PPA Modeling with Transfer Learning2021 ACM/IEEE 3rd Workshop on Machine Learning for CAD (MLCAD)10.1109/MLCAD52597.2021.9531109(1-6)Online publication date: 30-Aug-2021
          • (2021)ScaleDNN: Data Movement Aware DNN Training on Multi-GPU2021 IEEE/ACM International Conference On Computer Aided Design (ICCAD)10.1109/ICCAD51958.2021.9643503(1-9)Online publication date: 1-Nov-2021
          • (2021)Novel AI based pre-silicon Performance estimation and validation of complex System-on-Chip2021 IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE)10.1109/CSDE53843.2021.9718369(1-6)Online publication date: 8-Dec-2021
          • (2020)Evaluation of Neural Network Models for Performance Prediction of Scientific Applications2020 IEEE REGION 10 CONFERENCE (TENCON)10.1109/TENCON50793.2020.9293788(426-431)Online publication date: 16-Nov-2020
          • (2020)Multivariate Performance and Power Prediction of Algorithms on Simulation-Based Hardware Models2020 19th International Symposium on Parallel and Distributed Computing (ISPDC)10.1109/ISPDC51135.2020.00029(150-157)Online publication date: Jul-2020
          • Show More Cited By

          View Options

          View options

          Get Access

          Login options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media