Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Software-Hardware Codesign for Efficient Neural Network Acceleration

Published: 01 March 2017 Publication History

Abstract

Designers making deep learning computing more efficient cannot rely solely on hardware. Incorporating software-optimization techniques such as model compression leads to significant power savings and performance improvement. This article provides an overview of DeePhi's technology flow, including compression, compilation, and hardware acceleration. Two accelerators, named Aristotle and Descartes, are designed to achieve extremely high energy efficiency for both client and datacenter applications with convolutional neural network and recurrent neural network, respectively.

Cited By

View all
  • (2023)Trusted Heterogeneous Disaggregated ArchitecturesProceedings of the 14th ACM SIGOPS Asia-Pacific Workshop on Systems10.1145/3609510.3609812(72-79)Online publication date: 24-Aug-2023
  • (2023)A Cryogenic Artificial Synapse based on Superconducting MemristorProceedings of the Great Lakes Symposium on VLSI 202310.1145/3583781.3590203(143-148)Online publication date: 5-Jun-2023
  • (2022)ScanflowExpert Systems with Applications: An International Journal10.1016/j.eswa.2022.117232202:COnline publication date: 15-Sep-2022
  • Show More Cited By

Index Terms

  1. Software-Hardware Codesign for Efficient Neural Network Acceleration
    Index terms have been assigned to the content through auto-classification.

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image IEEE Micro
    IEEE Micro  Volume 37, Issue 2
    March 2017
    102 pages

    Publisher

    IEEE Computer Society Press

    Washington, DC, United States

    Publication History

    Published: 01 March 2017

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 28 Sep 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2023)Trusted Heterogeneous Disaggregated ArchitecturesProceedings of the 14th ACM SIGOPS Asia-Pacific Workshop on Systems10.1145/3609510.3609812(72-79)Online publication date: 24-Aug-2023
    • (2023)A Cryogenic Artificial Synapse based on Superconducting MemristorProceedings of the Great Lakes Symposium on VLSI 202310.1145/3583781.3590203(143-148)Online publication date: 5-Jun-2023
    • (2022)ScanflowExpert Systems with Applications: An International Journal10.1016/j.eswa.2022.117232202:COnline publication date: 15-Sep-2022
    • (2022)High-efficient MPSoC-based CNNs accelerator with optimized storage and dataflowThe Journal of Supercomputing10.1007/s11227-021-03909-y78:3(3205-3225)Online publication date: 1-Feb-2022
    • (2022)Towards An FPGA-targeted Hardware/Software Co-design Framework for CNN-based Edge ComputingMobile Networks and Applications10.1007/s11036-022-01985-927:5(2024-2035)Online publication date: 1-Oct-2022
    • (2021)Low-precision Floating-point Arithmetic for High-performance FPGA-based CNN AccelerationACM Transactions on Reconfigurable Technology and Systems10.1145/347459715:1(1-21)Online publication date: 9-Nov-2021
    • (2021)Uncertainty-aware Decisions in Cloud ComputingACM Computing Surveys10.1145/344758354:4(1-30)Online publication date: 24-May-2021
    • (2021)Warehouse-scale video acceleration: co-design and deployment in the wildProceedings of the 26th ACM International Conference on Architectural Support for Programming Languages and Operating Systems10.1145/3445814.3446723(600-615)Online publication date: 19-Apr-2021
    • (2021)FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with Fractional ActivationsThe 2021 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays10.1145/3431920.3439296(171-182)Online publication date: 17-Feb-2021
    • (2020)Exploiting the Relationship between Pruning Ratio and Compression Effect for Neural Network Model Based on TensorFlowSecurity and Communication Networks10.1155/2020/52186122020Online publication date: 1-Jan-2020
    • Show More Cited By

    View Options

    View options

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media