Nothing Special   »   [go: up one dir, main page]

Mohaidat et al., 2024 - Google Patents

A survey on neural network hardware accelerators

Mohaidat et al., 2024

Document ID
13726637778778386183
Author
Mohaidat T
Khalil K
Publication year
Publication venue
IEEE Transactions on Artificial Intelligence

External Links

Snippet

Artificial intelligence hardware accelerator is an emerging research for several applications and domains. The hardware accelerator's direction is to provide high computational speed with retaining low-cost and high learning performance. The main challenge is to design …
Continue reading at ieeexplore.ieee.org (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • G06F17/5009Computer-aided design using simulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • G06F17/5045Circuit design
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor; File system structures therefor
    • G06F17/30286Information retrieval; Database structures therefor; File system structures therefor in structured data stores
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers
    • G06F15/78Architectures of general purpose stored programme computers comprising a single central processing unit
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers
    • G06F15/80Architectures of general purpose stored programme computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
    • G06F15/8007Architectures of general purpose stored programme computers comprising an array of processing units with common control, e.g. single instruction multiple data processors single instruction multiple data [SIMD] multiprocessors
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for programme control, e.g. control unit
    • G06F9/06Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/12Computer systems based on biological models using genetic models
    • G06N3/126Genetic algorithms, i.e. information processing using digital simulations of the genetic system
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • G06N99/005Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F19/00Digital computing or data processing equipment or methods, specially adapted for specific applications
    • G06F19/10Bioinformatics, i.e. methods or systems for genetic or protein-related data processing in computational molecular biology
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a programme unit and a register, e.g. for a simultaneous processing of several programmes
    • G06F15/163Interprocessor communication
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F2217/00Indexing scheme relating to computer aided design [CAD]
    • G06F2217/78Power analysis and optimization
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computer systems utilising knowledge based models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL

Similar Documents

Publication Publication Date Title
Shuvo et al. Efficient acceleration of deep learning inference on resource-constrained edge devices: A review
Dave et al. Hardware acceleration of sparse and irregular tensor computations of ml models: A survey and insights
Blott et al. FINN-R: An end-to-end deep-learning framework for fast exploration of quantized neural networks
Conti et al. XNOR neural engine: A hardware accelerator IP for 21.6-fJ/op binary neural network inference
Mazumder et al. A survey on the optimization of neural network accelerators for micro-ai on-device inference
Mohaidat et al. A survey on neural network hardware accelerators
Daghero et al. Energy-efficient deep learning inference on edge devices
Zhang et al. Implementation of DNNs on IoT devices
Alawad Scalable FPGA accelerator for deep convolutional neural networks with stochastic streaming
Zhou et al. Addressing sparsity in deep neural networks
Sateesan et al. A survey of algorithmic and hardware optimization techniques for vision convolutional neural networks on FPGAs
Shi et al. Versagnn: a versatile accelerator for graph neural networks
Yan et al. FPGAN: an FPGA accelerator for graph attention networks with software and hardware co-optimization
de Prado et al. Automated design space exploration for optimized deployment of dnn on arm cortex-a cpus
Li et al. A survey of field programmable gate array (FPGA)-based graph convolutional neural network accelerators: challenges and opportunities
Qi et al. TRIM: A Design Space Exploration Model for Deep Neural Networks Inference and Training Accelerators
Lu et al. SparseNN: A performance-efficient accelerator for large-scale sparse neural networks
Liu et al. Review of neural network model acceleration techniques based on FPGA platforms
Shakibhamedan et al. Ease: Energy optimization through adaptation–a review of runtime energy-aware approximate deep learning algorithms
Zhang et al. Compilation and Optimizations for Efficient Machine Learning on Embedded Systems
Sun et al. Computation on sparse neural networks: an inspiration for future hardware
Wang et al. Balancing memory-accessing and computing over sparse DNN accelerator via efficient data packaging
Sun et al. Computation on sparse neural networks and its implications for future hardware
Langroudi et al. Digital neuromorphic chips for deep learning inference: a comprehensive study
Kao Performance-oriented FPGA-based convolution neural network designs