Nothing Special   »   [go: up one dir, main page]

Dong et al., 2021 - Google Patents

Hao: Hardware-aware neural architecture optimization for efficient inference

Dong et al., 2021

View PDF
Document ID
3662679487013120128
Author
Dong Z
Gao Y
Huang Q
Wawrzynek J
So H
Keutzer K
Publication year
Publication venue
2021 IEEE 29th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM)

External Links

Snippet

Automatic algorithm-hardware co-design for DNN has shown great success in improving the performance of DNNs on FPGAs. However, this process remains challenging due to the intractable search space of neural network architectures and hardware accelerator …
Continue reading at arxiv.org (PDF) (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • G06F17/5009Computer-aided design using simulation
    • G06F17/5018Computer-aided design using simulation using finite difference methods or finite element methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • G06F17/5045Circuit design
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/11Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/544Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices for evaluating functions by calculation
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/52Multiplying; Dividing
    • G06F7/523Multiplying only
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F2217/00Indexing scheme relating to computer aided design [CAD]
    • G06F2217/78Power analysis and optimization
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for programme control, e.g. control unit
    • G06F9/06Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06QDATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation, e.g. linear programming, "travelling salesman problem" or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F19/00Digital computing or data processing equipment or methods, specially adapted for specific applications
    • G06F19/10Bioinformatics, i.e. methods or systems for genetic or protein-related data processing in computational molecular biology
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers

Similar Documents

Publication Publication Date Title
Dong et al. Hao: Hardware-aware neural architecture optimization for efficient inference
Blott et al. FINN-R: An end-to-end deep-learning framework for fast exploration of quantized neural networks
Su et al. Redundancy-reduced mobilenet acceleration on reconfigurable logic for imagenet classification
Xu et al. AutoDNNchip: An automated DNN chip predictor and builder for both FPGAs and ASICs
US10656962B2 (en) Accelerate deep neural network in an FPGA
Yang et al. A fully onchip binarized convolutional neural network fpga impelmentation with accurate inference
Li et al. FLASH: F ast Neura l A rchitecture S earch with H ardware Optimization
Ke et al. Nnest: Early-stage design space exploration tool for neural network inference accelerators
Tao et al. Challenges in energy-efficient deep neural network training with FPGA
Reggiani et al. Pareto optimal design space exploration for accelerated CNN on FPGA
Ayachi et al. Optimizing neural networks for efficient FPGA implementation: A survey
Jahanshahi Tinycnn: A tiny modular CNN accelerator for embedded FPGA
Gong et al. N3h-core: Neuron-designed neural network accelerator via fpga-based heterogeneous computing cores
Vo et al. A deep learning accelerator based on a streaming architecture for binary neural networks
Sateesan et al. A survey of algorithmic and hardware optimization techniques for vision convolutional neural networks on FPGAs
Baharani et al. Deepdive: An integrative algorithm/architecture co-design for deep separable convolutional neural networks
Reggiani et al. Mix-gemm: An efficient hw-sw architecture for mixed-precision quantized deep neural networks inference on edge devices
Sivasankari et al. High-throughput and power-efficient convolutional neural network using one-pass processing elements
Shawahna et al. FxP-QNet: a post-training quantizer for the design of mixed low-precision DNNs with dynamic fixed-point representation
Mazouz et al. Automated CNN back-propagation pipeline generation for FPGA online training
Garbay et al. Cnn inference costs estimation on microcontrollers: the est primitive-based model
Liu et al. An efficient fpga-based depthwise separable convolutional neural network accelerator with hardware pruning
Sait et al. Optimization of FPGA-based CNN accelerators using metaheuristics
Kulkarni et al. Hybrid optimization for DNN model compression and inference acceleration
León-Vega et al. Automatic Generation of Resource and Accuracy Configurable Processing Elements