Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

NeuroSim: A Circuit-Level Macro Model for Benchmarking Neuro-Inspired Architectures in Online Learning

Published: 01 December 2018 Publication History

Abstract

Neuro-inspired architectures based on synaptic memory arrays have been proposed for on-chip acceleration of weighted sum and weight update in machine/deep learning algorithms. In this paper, we developed NeuroSim, a circuit-level macro model that estimates the area, latency, dynamic energy, and leakage power to facilitate the design space exploration of neuro-inspired architectures with mainstream and emerging device technologies. NeuroSim provides flexible interface and a wide variety of design options at the circuit and device level. Therefore, NeuroSim can be used by neural networks (NNs) as a supporting tool to provide circuit-level performance evaluation. With NeuroSim, an integrated framework can be built with hierarchical organization from the device level (synaptic device properties) to the circuit level (array architectures) and then to the algorithm level (NN topology), enabling instruction-accurate evaluation on the learning accuracy as well as the circuit-level performance metrics at the run-time of online learning. Using multilayer perceptron as a case-study algorithm, we investigated the impact of the “analog” emerging nonvolatile memory (eNVM)’s “nonideal” device properties and benchmarked the tradeoffs between SRAM, digital, and analog eNVM-based architectures for online learning and offline classification.

References

[1]
J. Hasler and B. Marr, “Finding a roadmap to achieve large neuromorphic hardware systems,” Front. Neurosci., vol. 7, no. 118, pp. 1–29, 2013.
[2]
J. Von Neumann, “The principles of large-scale computing machines,” Ann. History Comput., vol. 3, no. 3, pp. 263–273, Jul./Sep. 1981.
[3]
R. Ananthanarayanan, S. K. Esser, H. D. Simon, and D. S. Modha, “The cat is out of the bag: Cortical simulations with 109 neurons, 1013 synapses,” in Proc. Conf. High Perform. Comput. Netw. Stor. Anal., Portland, OR, USA, 2009, pp. 1–12.
[4]
S. Schmittet al., “Neuromorphic hardware in the loop: Training a deep spiking network on the brainscales wafer-scale system,” in Proc. Int. Joint Conf. Neural Netw. (IJCNN), 2017, pp. 2227–2234.
[5]
B. V. Benjaminet al., “Neurogrid: A mixed-analog–digital multichip system for large-scale neural simulations,” Proc. IEEE, vol. 102, no. 5, pp. 699–716, May 2014.
[6]
P. A. Merollaet al., “A million spiking-neuron integrated circuit with a scalable communication network and interface,” Science, vol. 345, no. 6197, pp. 668–673, 2014.
[7]
Y.-H. Chen, T. Krishna, J. Emer, and V. Sze, “14.5 eyeriss: An energy-efficient reconfigurable accelerator for deep convolutional neural networks,” in Proc. IEEE Int. Solid-State Circuits Conf. (ISSCC), San Francisco, CA, USA, 2016, pp. 262–263.
[8]
D. Shin, J. Lee, J. Lee, and H.-J. Yoo, “14.2 DNPU: An 8.1TOPS/W reconfigurable CNN-RNN processor for general-purpose deep neural networks,” in Proc. Int. Solid-State Circuits Conf. (ISSCC), San Francisco, CA, USA, 2017, pp. 240–241.
[9]
S. Yu, Ed., Neuro-Inspired Computing Using Resistive Synaptic Devices. Cham, Switzerland: Springer, 2017.
[10]
P.-Y. Chenet al., “Mitigating effects of non-ideal synaptic device characteristics for on-chip learning,” in Proc. ACM/IEEE Int. Conf. Comput.-Aided Design (ICCAD), Austin, TX, USA, 2015, pp. 194–199.
[11]
G. W. Burret al., “Experimental demonstration and tolerancing of a large-scale neural network (165 000 synapses) using phase-change memory as the synaptic weight element,” IEEE Trans. Electron Devices, vol. 62, no. 11, pp. 3498–3507, Nov. 2015.
[12]
S. Agarwalet al., “Resistive memory device requirements for a neural algorithm accelerator,” in Proc. Int. Joint Conf. Neural Netw. (IJCNN), Vancouver, BC, Canada, 2016, pp. 929–938.
[13]
P. Chiet al., “PRIME: A novel processing-in-memory architecture for neural network computation in ReRAM-based main memory,” in Proc. ACM/IEEE Int. Symp. Comput. Archit. (ISCA), Seoul, South Korea, 2016, pp. 27–39.
[14]
X. Liuet al., “Harmonica: A framework of heterogeneous computing systems with memristor-based neuromorphic computing accelerators,” IEEE Trans. Circuits Syst. I, Reg. Papers, vol. 63, no. 5, pp. 617–628, May 2016.
[15]
A. Shafieeet al., “ISAAC: A convolutional neural network accelerator with in-situ analog arithmetic in crossbars,” in Proc. Int. Symp. Comput. Archit. (ISCA), Seoul, South Korea, 2016, pp. 14–26.
[16]
L. Xiaet al., “MNSIM: Simulation platform for memristor-based neuromorphic computing system,” in Proc. ACM/IEEE Design Autom. Test Europe Conf. Exhibit. (DATE), Dresden, Germany, 2016, pp. 469–474.
[17]
MLP Simlator (+NeuroSim) Version 1.0. [Online]. Available: https://github.com/neurosim/MLP_NeuroSim
[18]
D. Kuzum, S. Yu, and H.-S. P. Wong, “Synaptic electronics: Materials, devices and applications,” Nanotechnology, vol. 24, no. 38, 2013, Art. no.
[19]
X. Guoet al., “Temperature-insensitive analog vector-by-matrix multiplier based on 55 nm NOR flash memory cells,” in Proc. IEEE Custom Integr. Circuits Conf. (CICC), Austin, TX, USA, 2017, pp. 1–4.
[20]
S. H. Joet al., “Nanoscale memristor device as synapse in neuromorphic systems,” Nano Lett., vol. 10, no. 4, pp. 1297–1301, 2010.
[21]
L. Gaoet al., “Fully parallel write/read in resistive synaptic array for accelerating on-chip learning,” Nanotechnology, vol. 26, no. 45, 2015, Art. no.
[22]
S. Parket al., “Neuromorphic speech systems using advanced ReRAM-based synapse,” in Proc. IEEE Int. Electron Device Meeting (IEDM), Washington, DC, USA, 2013, pp. 625–628.
[23]
J. Wooet al., “Improved synaptic behavior under identical pulses using AlOx/HfO2 bilayer RRAM array for neuromorphic systems,” IEEE Electron Device Lett., vol. 37, no. 8, pp. 994–997, Aug. 2016.
[24]
S. Yuet al., “A low energy oxide-based electronic synaptic device for neuromorphic visual systems with tolerance to device variation,” Adv. Mater., vol. 25, no. 12, pp. 1774–1779, 2013.
[25]
M. Preziosoet al., “Training and operation of an integrated neuromorphic network based on metal-oxide memristors,” Nature, vol. 521, no. 7550, pp. 61–64, 2015.
[26]
D. Kuzum, R. G. D. Jeyasingh, B. Lee, and H.-S. P. Wong, “Nanoelectronic programmable synapses based on phase change materials for brain-inspired computing,” Nano Lett., vol. 12, no. 5, pp. 2179–2186, 2011.
[27]
M. Suriet al., “Phase change memory as synapse for ultra-dense neuromorphic systems: Application to complex visual pattern extraction,” in Proc. IEEE Int. Electron Devices Meeting (IEDM), Washington, DC, USA, 2011, pp. 79–82.
[28]
M. Hu, H. Li, Q. Wu, and G. S. Rose, “Hardware realization of BSB recall function using memristor crossbar arrays,” in Proc. Design Autom. Conf. (DAC), San Francisco, CA, USA, 2012, pp. 498–503.
[29]
P.-Y. Chenet al., “Technology-design co-optimization of resistive cross-point array for accelerating learning algorithms on chip,” in Proc. Design Autom. Test Europe Conf. Exhibit. (DATE), Grenoble, France, 2015, pp. 854–859.
[30]
J. Liang and H.-S. P. Wong, “Cross-point memory array without cell selectors—Device characteristics and data storage pattern dependencies,” IEEE Trans. Electron Devices, vol. 57, no. 10, pp. 2531–2538, Oct. 2010.
[31]
P.-Y. Chen, L. Gao, and S. Yu, “Design of resistive synaptic array for implementing on-chip sparse learning,” IEEE Trans. Multi-Scale Comput. Syst., vol. 2, no. 4, pp. 257–264, Oct./Dec. 2016.
[32]
S. Yu and P.-Y. Chen, “Emerging memory technologies: Recent trends and prospects,” IEEE Solid State Circuits Mag., vol. 8, no. 2, pp. 43–56, Jun. 2016.
[33]
S. Yuet al., “Scaling-up resistive synaptic arrays for neuro-inspired architecture: Challenges and prospect,” in Proc. IEEE Int. Electron Devices Meeting (IEDM), Washington, DC, USA, 2015, pp. 451–454.
[34]
S. J. E. Wilton and N. P. Jouppi, “CACTI: An enhanced cache access and cycle time model,” IEEE J. Solid-State Circuits, vol. 31, no. 5, pp. 677–688, May 1996.
[35]
X. Dong, C. Xu, Y. Xie, and N. P. Jouppi, “NVSim: A circuit-level performance, energy, and area model for emerging nonvolatile memory,” IEEE Trans. Comput.-Aided Design Integr. Circuits Syst., vol. 31, no. 7, pp. 994–1007, Jul. 2012.
[36]
D. Kadetotadet al., “Parallel architecture with resistive crosspoint array for dictionary learning acceleration,” IEEE J. Emerg. Sel. Topic Circuits Syst., vol. 5, no. 2, pp. 194–204, Jun. 2015.
[37]
Predictive Technology Model (PTM). [Online]. Available: http://ptm.asu.edu/
[38]
W. Qianet al., “Energy-efficient adaptive computing with multifunctional memory,” IEEE Trans. Circuits Syst. II, Exp. Briefs, vol. 64, no. 2, pp. 191–195, Feb. 2017.
[39]
P.-Y. Chen and S. Yu, “Partition SRAM and RRAM based synaptic arrays for neuro-inspired computing,” in Proc. IEEE Int. Symp. Circuits Syst. (ISCAS), Montreal, QC, Canada, 2016, pp. 2310–2313.
[40]
P. Rosenfeld, E. Cooper-Balis, and B. Jacob, “DRAMSim2: A cycle accurate memory system simulator,” IEEE Comput. Archit. Lett., vol. 10, no. 1, pp. 16–19, Jan./Jun. 2011.
[42]
Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278–2324, Nov. 1998.
[43]
T. Tang, L. Xia, B. Li, Y. Wang, and H. Yang, “Binary convolutional neural network on RRAM,” in Proc. ACM/IEEE Asia South Pac. Design Autom. Conf. (ASP-DAC), 2017, pp. 782–787.
[44]
L. Gao, P.-Y. Chen, and S. Yu, “Programming protocol optimization for analog weight tuning in resistive memories,” IEEE Electron Device Lett., vol. 36, no. 11, pp. 1157–1159, Nov. 2015.

Cited By

View all
  • (2025)HyperMR: Efficient Hypergraph-enhanced Matrix Storage on Compute-in-Memory ArchitectureProceedings of the ACM on Management of Data10.1145/37096953:1(1-27)Online publication date: 11-Feb-2025
  • (2025)A ReRAM-Based Processing-In-Memory Architecture for Hyperdimensional ComputingIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2024.344581244:2(512-524)Online publication date: 1-Feb-2025
  • (2025)Light-CIM: A Lightweight ADC/DAC-Fewer RRAM CIM DNN Accelerator With Fully Analog Tiles and Nonideality-Aware Algorithm for Consumer ElectronicsIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2024.343569044:2(602-612)Online publication date: 1-Feb-2025
  • Show More Cited By

Index Terms

  1. NeuroSim: A Circuit-Level Macro Model for Benchmarking Neuro-Inspired Architectures in Online Learning
            Index terms have been assigned to the content through auto-classification.

            Recommendations

            Comments

            Please enable JavaScript to view thecomments powered by Disqus.

            Information & Contributors

            Information

            Published In

            cover image IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
            IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems  Volume 37, Issue 12
            Dec. 2018
            254 pages

            Publisher

            IEEE Press

            Publication History

            Published: 01 December 2018

            Qualifiers

            • Research-article

            Contributors

            Other Metrics

            Bibliometrics & Citations

            Bibliometrics

            Article Metrics

            • Downloads (Last 12 months)0
            • Downloads (Last 6 weeks)0
            Reflects downloads up to 05 Mar 2025

            Other Metrics

            Citations

            Cited By

            View all
            • (2025)HyperMR: Efficient Hypergraph-enhanced Matrix Storage on Compute-in-Memory ArchitectureProceedings of the ACM on Management of Data10.1145/37096953:1(1-27)Online publication date: 11-Feb-2025
            • (2025)A ReRAM-Based Processing-In-Memory Architecture for Hyperdimensional ComputingIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2024.344581244:2(512-524)Online publication date: 1-Feb-2025
            • (2025)Light-CIM: A Lightweight ADC/DAC-Fewer RRAM CIM DNN Accelerator With Fully Analog Tiles and Nonideality-Aware Algorithm for Consumer ElectronicsIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2024.343569044:2(602-612)Online publication date: 1-Feb-2025
            • (2025)PINSim: A Processing In- and Near-Sensor Simulator to Model Intelligent Vision SensorsIEEE Computer Architecture Letters10.1109/LCA.2024.352277724:1(17-20)Online publication date: 1-Jan-2025
            • (2024)A Cascaded ReRAM-based Crossbar Architecture for Transformer Neural Network AccelerationACM Transactions on Design Automation of Electronic Systems10.1145/370103430:1(1-23)Online publication date: 18-Oct-2024
            • (2024)Systolic Array Acceleration of Spiking Neural Networks with Application-Independent Split-Time Temporal CodingProceedings of the 29th ACM/IEEE International Symposium on Low Power Electronics and Design10.1145/3665314.3672292(1-6)Online publication date: 5-Aug-2024
            • (2024)ReHarvest: An ADC Resource-Harvesting Crossbar Architecture for ReRAM-Based DNN AcceleratorsACM Transactions on Architecture and Code Optimization10.1145/365920821:3(1-26)Online publication date: 17-Apr-2024
            • (2024)NavCim: Comprehensive Design Space Exploration for Analog Computing-in-Memory ArchitecturesProceedings of the 2024 International Conference on Parallel Architectures and Compilation Techniques10.1145/3656019.3676946(168-182)Online publication date: 14-Oct-2024
            • (2024)MixMixQ: Quantization with Mixed Bit-Sparsity and Mixed Bit-Width for CIM AcceleratorsProceedings of the Great Lakes Symposium on VLSI 202410.1145/3649476.3658809(537-540)Online publication date: 12-Jun-2024
            • (2024)ReCG: ReRAM-Accelerated Sparse Conjugate GradientProceedings of the 61st ACM/IEEE Design Automation Conference10.1145/3649329.3656515(1-6)Online publication date: 23-Jun-2024
            • Show More Cited By

            View Options

            View options

            Figures

            Tables

            Media

            Share

            Share

            Share this Publication link

            Share on social media