Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article
Free access
Just Accepted

A Design Flow for Scheduling Spiking Deep Convolutional Neural Networks on Heterogeneous Neuromorphic System-on-Chip

Online AM: 02 December 2023 Publication History

Abstract

Neuromorphic systems-on-chip (NSoCs) integrate CPU cores and neuromorphic hardware accelerators on the same chip. These platforms can execute spiking deep convolutional neural networks (SDCNNs) with a low energy footprint. Modern NSoCs are heterogeneous in terms of their computing, communication, and storage resources. This makes scheduling SDCNN operations a combinatorial problem of exploring an exponentially-large state space in determining mapping, ordering, and timing of operations to achieve a target hardware performance, e.g., throughput.
We propose a systematic design flow to schedule SDCNNs on an NSoC. Our scheduler, called SMART (SDCNN MApping, OrdeRing, and Timing), branches the combinatorial optimization problem into computationally-relaxed sub-problems that generate fast solutions without significantly compromising the solution quality. SMART improves performance by efficiently incorporating the heterogeneity in computing, communication, and storage resources. SMART operates in four steps. First, it creates a self-timed execution schedule to map operations to compute resources, maximizing throughput. Second, it uses an optimization strategy to distribute activation and synaptic weights to storage resources, minimizing data communication-related overhead. Third, it constructs an inter-processor communication (IPC) graph with a transaction order for its communication actors. This transaction order is created using a transaction partial order algorithm, which minimizes contention on the shared communication resources. Finally, it schedules this IPC graph to hardware by overlapping communication with the computation, and leveraging operation, pipeline, and batch parallelism.
We evaluate SMART using 10 representative image, object, and language-based SDCNNs. Results show that SMART increases throughput by an average 23%, compared to a state-of-the-art scheduler. SMART is implemented entirely in software as a compiler extension. It doesn’t require any change in a neuromorphic hardware or its interface to CPUs. It improves throughput with only a marginal increase in the compilation time. SMART is released under the open-source MIT licensing at https://github.com/drexel-DISCO/SMART to foster future research.

References

[1]
Mayez A. Al-Mouhamed. 1990. Lower bound on the number of processors and time for scheduling precedence graphs with communication costs. IEEE Transactions on Software Engineering 16, 12 (1990), 1390–1401.
[2]
Daniel Auge, Julian Hille, Etienne Mueller, and Alois Knoll. 2021. A survey of encoding techniques for signal processing in spiking neural networks. Neural Processing Letters 53, 6 (2021), 4693–4710.
[3]
Adarsha Balaji, Prathyusha Adiraju, Hirak J Kashyap, Anup Das, Jeffrey L Krichmar, Nikil D Dutt, and Francky Catthoor. 2020. PyCARL: A PyNN interface for hardware-software co-simulation of spiking neural network. In International Joint Conference on Neural Networks (IJCNN).
[4]
Adarsha Balaji, Federico Corradi, Anup Das, Sandeep Pande, Siebren Schaafsma, and Francky Catthoor. 2018. Power-accuracy trade-offs for heartbeat classification on neural networks hardware. Journal of Low Power Electronics 14, 4 (2018), 508–519.
[5]
Adarsha Balaji and Anup Das. 2019. A Framework for the Analysis of Throughput-Constraints of SNNs on Neuromorphic Hardware. In IEEE Annual Symposium on VLSI (ISVLSI).
[6]
Adarsha Balaji, Anup Das, Yuefeng Wu, Khanh Huynh, Francesco G Dell’Anna, Giacomo Indiveri, Jeffrey L Krichmar, Nikil D Dutt, Siebren Schaafsma, and Francky Catthoor. 2019. Mapping spiking neural networks to neuromorphic hardware. IEEE Transactions on Very Large Scale Integration (VLSI) Systems 28, 1(2019), 76–86.
[7]
Adarsha Balaji, Phu Khanh Huynh, Francky Catthoor, Nikil D Dutt, Jeffrey L Krichmar, and Anup Das. 2023. NeuSB: A scalable interconnect architecture for spiking neuromorphic hardware. IEEE Transactions on Emerging Topics in Computing (2023).
[8]
Adarsha Balaji, Shihao Song, Anup Das, Jeffrey Krichmar, Nikil Dutt, James Shackleford, Nagarajan Kandasamy, and Francky Catthoor. 2020. Enabling resource-aware mapping of spiking neural networks via spatial decomposition. Embedded Systems Letters 13, 3 (2020), 142–145.
[9]
Adarsha Balaji, Shihao Song, Twisha Titirsha, Anup Das, Jeffrey Krichmar, Nikil Dutt, James Shackleford, Nagarajan Kandasamy, and Francky Catthoor. 2021. NeuroXplorer 1.0: An Extensible Framework for Architectural Exploration with Spiking Neural Networks. In International Conference on Neuromorphic Systems (ICONS).
[10]
Adarsha Balaji, Yuefeng Wu, Anup Das, Francky Catthoor, and Siebren Schaafsma. 2019. Exploration of segmented bus as scalable global interconnect for neuromorphic computing. In Great Lakes Symposium on VLSI (GLSVLSI).
[11]
Neal Bambha, Vida Kianzad, Mukul Khandelia, and Shuvra S Bhattacharyya. 2002. Intermediate representations for design automation of multiprocessor DSP systems. 7, 4 (2002), 307–323.
[12]
Rajeshwari Banakar, Stefan Steinke, Bo-Sik Lee, Mahesh Balakrishnan, and Peter Marwedel. 2002. Scratchpad memory: design alternative for cache on-chip memory in embedded systems. In Proceedings of the tenth international symposium on Hardware/software codesign. 73–78.
[13]
Richard H Bartels and Gene H Golub. 1969. The simplex method of linear programming using LU decomposition. Commun. ACM 12, 5 (1969), 266–268.
[14]
Alan Bawden. 1986. Connection graphs. In ACM conference on LISP and functional programming. 258–265.
[15]
Luca Benini and Giovanni De Micheli. 2002. Networks on chip: A new paradigm for systems on chip design. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition. IEEE, 418–419.
[16]
Shuvra S Bhattacharyya, Praveen K Murthy, and Edward A Lee. 1996. Software synthesis from dataflow graphs. Vol.  360. Springer Science & Business Media.
[17]
Jacek Blazewicz. 1976. Scheduling dependent tasks with different arrival times to meet deadlines. In Proceedings of the International Workshop organized by the Commision of the European Communities on Modelling and Performance Evaluation of Computer Systems. 57–65.
[18]
Christian Bliek1ú, Pierre Bonami, and Andrea Lodi. 2014. Solving mixed-integer quadratic programming problems with IBM-CPLEX: a progress report. In Proceedings of the twenty-sixth RAMP symposium. 16–17.
[19]
BrainChip. 2022. Akida Neuromorphic System- on-Chip. https://www.brainchipinc.com/ products/akida-neuromorphic-system-on-chip
[20]
BrainChip. 2022. MetaTF development environment. https://brainchipinc.com/metatf-development-environment/
[21]
Yongqiang Cao, Yang Chen, and Deepak Khosla. 2015. Spiking deep convolutional neural networks for energy-efficient object recognition. International Journal of Computer Vision 113 (2015), 54–66.
[22]
Francky Catthoor, Matthias Hartmann, Jose Ignacio Gomez, Christian Tenllado, Sotiris Xydis, Javier Setoain Rodrigo, Thomas Papastergiou, Christos Baloukas, Anup Kumar Das, Dimitrios Soudris, et al. 2020. Memory structure comprising scratchpad memory. US Patent 10,592,430.
[23]
Francky Catthoor, Srinjoy Mitra, Anup Das, and Siebren Schaafsma. 2018. Very large-scale neuromorphic systems for biological signal processing. In CMOS Circuits for Biological Sensing and Processing.
[24]
Yunji Chen, Tianshi Chen, Zhiwei Xu, Ninghui Sun, and Olivier Temam. 2016. DianNao family: energy-efficient hardware accelerators for machine learning. Commun. ACM 59, 11 (2016), 105–112.
[25]
Yu-Hsin Chen, Tien-Ju Yang, et al. 2019. Eyeriss v2: A flexible accelerator for emerging deep neural networks on mobile devices. IEEE Journal on Emerging and Selected Topics in Circuits and Systems (JETCAS) 9, 2(2019), 292–308.
[26]
Serena Curzel, Nicolas Bohm Agostini, Shihao Song, Ismet Dagli, Ankur Limaye, Cheng Tan, Marco Minutoli, Vito Giovanni Castellana, Vinay Amatya, Joseph Manzano, et al. 2021. Automated Generation of Integrated Digital and Spiking Neuromorphic Machine Learning Accelerators. In International Conference on Computer-Aided Design (ICCAD).
[27]
Morteza Damavandpeyma, Sander Stuijk, Twan Basten, Marc Geilen, and Henk Corporaal. 2012. Modeling static-order schedules in synchronous dataflow graphs. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition. IEEE, 775–780.
[28]
Anup Das. 2022. Real-Time Scheduling of Machine Learning Operations on Heterogeneous Neuromorphic SoC. In 2022 20th ACM/IEEE International Conference on Formal Methods and Models for System Design (MEMOCODE). IEEE, 1–12.
[29]
Anup Das. 2023. Preserving Privacy of Neuromorphic Hardware From PCIe Congestion Side-Channel Attack. In IEEE Annual International Computer Software and Applications Conference (COMPSAC).
[30]
Anup Das, Francky Catthoor, and Siebren Schaafsma. 2018. Heartbeat classification in wearables using multi-layer perceptron and time-frequency joint distribution of ECG. In International conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE). 69–74.
[31]
Anup Das and Akash Kumar. 2018. Dataflow-Based Mapping of Spiking Neural Networks on Neuromorphic Hardware. In Great Lakes Symposium on VLSI (GLSVLSI).
[32]
Anup Das, Akash Kumar, and Bharadwaj Veeravalli. 2014. Energy-aware task mapping and scheduling for reliable embedded computing systems. ACM Transactions on Embedded Computing Systems 13, 2s (2014), 1–27.
[33]
Anup Das, Akash Kumar, and Bharadwaj Veeravalli. 2014. Temperature aware energy-reliability trade-offs for mapping of throughput-constrained applications on multimedia MPSoCs. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition.
[34]
Anup Das, Akash Kumar, and Bharadwaj Veeravalli. 2015. Reliability and energy-aware mapping and scheduling of multimedia applications on multiprocessor systems. IEEE Transactions on Parallel and Distributed Systems 27, 3 (2015), 869–884.
[35]
Anup Das, Akash Kumar, and Bharadwaj Veeravalli. 2015. Reliability and energy-aware mapping and scheduling of multimedia applications on multiprocessor systems. IEEE Transactions on Parallel and Distributed Systems 27, 3 (2015), 869–884.
[36]
Anup Das, Yuefeng Wu, Khanh Huynh, Francesco Dell’Anna, Francky Catthoor, and Siebren Schaafsma. 2018. Mapping of local and global synapses on spiking neuromorphic hardware. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition.
[37]
Mike Davies, Narayan Srinivasa, Tsung-Han Lin, Gautham Chinya, Yongqiang Cao, Sri Harsha Choday, Georgios Dimou, Prasad Joshi, Nabil Imam, Shweta Jain, et al. 2018. Loihi: A neuromorphic manycore processor with on-chip learning. Ieee Micro 38, 1 (2018), 82–99.
[38]
Michael V DeBole, Brian Taba, Arnon Amir, Filipp Akopyan, Alexander Andreopoulos, William P Risk, Jeff Kusnitz, Carlos Ortega Otero, Tapan K Nayak, Rathinakumar Appuswamy, et al. 2019. TrueNorth: Accelerating from zero to 64 million neurons in 10 years. Computer 52, 5 (2019), 20–29.
[39]
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In IEEE conference on computer vision and pattern recognition. Ieee, 248–255.
[40]
Steve Furber, Francesco Galluppi, Steve Temple, and Luis A Plana. 2014. The SpiNNaker project. Proc. IEEE 102, 5 (2014), 652–665.
[41]
Francesco Galluppi, Sergio Davies, Alexander Rast, Thomas Sharp, Luis A Plana, and Steve Furber. 2012. A hierachical configuration system for a massively parallel neural hardware platform. In International Conference on Computing Frontiers (CF).
[42]
Amir Hossein Ghamarian, MCW Geilen, Twan Basten, Bart D Theelen, Mohammad Reza Mousavi, and Sander Stuijk. 2006. Liveness and boundedness of synchronous data flow graphs. In Formal Methods in Computer Aided Design. IEEE, 68–75.
[43]
Amir Hossein Ghamarian, Marc CW Geilen, Sander Stuijk, Twan Basten, Bart D Theelen, Mohammad Reza Mousavi, Arno JM Moonen, and Marco JG Bekooij. 2006. Throughput analysis of synchronous data flow graphs. In International Conference on Application of Concurrency to System Design (ACSD).
[44]
Dorit S Hochba. 1997. Approximation algorithms for NP-hard problems. ACM Sigact News 28, 2 (1997), 40–52.
[45]
Alexandre Honorat, Karol Desnos, Mickaël Dardaillon, and Jean-François Nezan. 2020. A fast heuristic to pipeline sdf graphs. In SAMOS. Springer, 139–151.
[46]
Phu Khanh Huynh, M Lakshmi Varshika, Ankita Paul, Murat Isik, Adarsha Balaji, and Anup Das. 2022. Implementing spiking neural networks on neuromorphic architectures: A review. arXiv preprint arXiv:2202.08897(2022).
[47]
Jing-Jang Hwang, Yuan-Chieh Chow, Frank D Anger, and Chung-Yee Lee. 1989. Scheduling precedence graphs in systems with interprocessor communication times. SIAM J. Comput. 18, 2 (1989), 244–257.
[48]
Giacomo Indiveri, Federico Corradi, and Ning Qiao. 2015. Neuromorphic architectures for spiking deep neural networks. In IEDM.
[49]
Dowhan Jeong, Jangryul Kim, Mari-Liis Oldja, and Soonhoi Ha. 2021. Parallel scheduling of multiple SDF graphs onto heterogeneous processors. IEEE Access 9(2021), 20493–20507.
[50]
Yu Ji, YouHui Zhang, ShuangChen Li, Ping Chi, CiHang Jiang, Peng Qu, Yuan Xie, and WenGuang Chen. 2016. NEUTRAMS: Neural network transformation and co-design under neuromorphic hardware constraints. In International Symposium on Microarchitecture (MICRO).
[51]
Li Jiashu, Anup Das, and Akash Kumar. 2012. A design flow for partially reconfigurable heterogeneous multi-processor platforms. In IEEE International Symposium on Rapid System Prototyping (RSP). IEEE, 170–176.
[52]
Narendra Karmarkar. 1984. A new polynomial-time algorithm for linear programming. In Proceedings of the sixteenth annual ACM symposium on Theory of computing. 302–311.
[53]
Yu-Kwong Kwok and Ishfaq Ahmad. 1999. Static scheduling algorithms for allocating directed task graphs to multiprocessors. ACM Computing Surveys (CSUR) 31, 4 (1999), 406–471.
[54]
GrAI Matter Labs. 2022. GrAI Chip and GrAIFlow Software. https://www.graimatterlabs.ai/product
[55]
Eugene L Lawler and David E Wood. 1966. Branch-and-bound methods: A survey. Operations research 14, 4 (1966), 699–719.
[56]
E.A. Lee and D.G. Messerschmitt. 1987. Synchronous data flow. Proc. IEEE 75, 9 (1987), 1235–1245.
[57]
Edward A Lee. 1991. Consistency in dataflow graphs. IEEE Transactions on Parallel and Distributed systems 2, 2 (1991), 223–235.
[58]
Edward Ashford Lee and David G Messerschmitt. 1987. Static scheduling of synchronous data flow programs for digital signal processing. IEEE Transactions on computers 100, 1 (1987), 24–35.
[59]
Chit-Kwan Lin, Andreas Wild, Gautham N. Chinya, Tsung-Han Lin, Mike Davies, and Hong Wang. 2018. Mapping Spiking Neural Networks onto a Manycore Neuromorphic Architecture. In Programming Language Design and Implementation (PLDI).
[60]
Qian Liu, Ole Richter, Carsten Nielsen, Sadique Sheik, Giacomo Indiveri, and Ning Qiao. 2019. Live demonstration: face recognition on an ultra-low power event-driven convolutional neural network ASIC. In CVPR Workshops.
[61]
Xiaoxiao Liu, Wei Wen, Xuehai Qian, Hai Li, and Yiran Chen. 2018. Neu-NoC: A high-efficient interconnection network for accelerated neuromorphic systems. In 2018 23rd Asia and South Pacific Design Automation Conference (ASP-DAC). IEEE, 141–146.
[62]
Wolfgang Maass. 1997. Networks of spiking neurons: the third generation of neural network models. Neural Networks 10, 9 (1997), 1659–1671.
[63]
Maxime Martinasso, Grzegorz Kwasniewski, Sadaf R Alam, Thomas C Schulthess, and Torsten Hoefler. 2016. A PCIe congestion-aware performance model for densely populated accelerator servers. In International Conference for High Performance Computing, Networking, Storage and Analysis.
[64]
Carver Mead. 1990. Neuromorphic electronic systems. Proc. IEEE 78, 10 (1990), 1629–1636.
[65]
Saber Moradi, Ning Qiao, Fabio Stefanini, and Giacomo Indiveri. 2017. A scalable multicore architecture with heterogeneous memory structures for dynamic neuromorphic asynchronous processors (DYNAPs). IEEE Transactions on Biomedical Circuits and Systems 12, 1 (2017), 106–122.
[66]
Lars Niedermeier, Kexin Chen, Jinwei Xing, Anup Das, Jeffrey Kopsick, Eric Scott, Nate Sutton, Killian Weber, Nikil Dutt, and Jeffrey L Krichmar. 2022. CARLsim 6: An open source library for large-scale, biologically detailed spiking neural network simulation. In International Joint Conference on Neural Networks (IJCNN).
[67]
Subhankar Pal, Swagath Venkataramani, Viji Srinivasan, and Kailash Gopalakrishnan. 2022. OnSRAM: Efficient Inter-Node On-Chip Scratchpad Management in Deep Learning Accelerators. ACM Transactions on Embedded Computing Systems (2022).
[68]
Ankita Paul, Shihao Song, Twisha Titirsha, and Anup Das. 2022. On the Mitigation of Read Disturbances in Neuromorphic Inference Hardware. IEEE Design & Test (2022).
[69]
Ankita Paul, Stefan Wagner, and Anup Das. 2022. Learning in feedback-driven recurrent spiking neural networks using full-force training. In International Joint Conference on Neural Networks (IJCNN).
[70]
Joo Pedro Pedroso. 2011. Optimization with gurobi and python. INESC Porto and Universidade do Porto, Porto, Portugal 1 (2011).
[71]
Larry L Peterson, Nick C Buchholz, and Richard D Schlichting. 1989. Preserving and using context information in interprocess communication. ACM Transactions on Computer Systems (TOCS) 7, 3 (1989), 217–246.
[72]
Andrei Radulescu and Arjan JC Van Gemund. 2002. Low-cost task scheduling for distributed-memory machines. IEEE transactions on parallel and distributed systems 13, 6 (2002), 648–658.
[73]
Kathrin Rosvall and Ingo Sander. 2014. A constraint-based design space exploration framework for real-time applications on MPSoCs. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition. IEEE.
[74]
Kaushik Roy, Akhilesh Jaiswal, and Priyadarshini Panda. 2019. Towards spike-based machine intelligence with neuromorphic computing. Nature 575, 7784 (2019), 607–617.
[75]
Abhronil Sengupta, Yuting Ye, Robert Wang, Chiao Liu, and Kaushik Roy. 2019. Going deeper in spiking neural networks: VGG and residual architectures. Frontiers in Neuroscience 13 (2019), 95.
[76]
Luping Shi, Jing Pei, Ning Deng, Dong Wang, Lei Deng, Yu Wang, Youhui Zhang, Feng Chen, Mingguo Zhao, Sen Song, et al. 2015. Development of a neuromorphic computing system. In International Electron Devices Meeting (IEDM).
[77]
Gilbert C Sih and Edward A Lee. 1993. A compile-time scheduling heuristic for interconnection-constrained heterogeneous processor architectures. IEEE transactions on Parallel and Distributed systems 4, 2 (1993), 175–187.
[78]
Abhishek Singh, Pontus Ekberg, and Sanjoy Baruah. 2019. Uniprocessor scheduling of real-time synchronous dataflow tasks. Real-Time Systems 55(2019), 1–31.
[79]
Amit Kumar Singh, Anup Das, and Akash Kumar. 2013. Energy optimization by exploiting execution slacks in streaming applications on multiprocessor systems.
[80]
Oliver Sinnen and Leonel A Sousa. 2005. Communication contention in task scheduling. IEEE Transactions on Parallel and Distributed Systems 16, 6 (2005), 503–515.
[81]
Shihao Song, Adarsha Balaji, Anup Das, and Nagarajan Kandasamy. 2022. Design-technology co-optimization for NVM-based neuromorphic processing elements. ACM Transactions on Embedded Computing Systems 21, 6 (2022), 1–27.
[82]
Shihao Song, Adarsha Balaji, Anup Das, Nagarajan Kandasamy, and James Shackleford. 2020. Compiling spiking neural networks to neuromorphic hardware. In International Conference on Languages, Compilers, and Tools for Embedded Systems (LCTES).
[83]
Shihao Song, Harry Chong, Adarsha Balaji, Anup Das, James Shackleford, and Nagarajan Kandasamy. 2021. DFSynthesizer: Dataflow-based synthesis of spiking neural networks to neuromorphic hardware. ACM Transactions on Embedded Computing Systems 2, 3 (2021), 1–4.
[84]
Shihao Song and Anup Das. 2020. A case for lifetime reliability-aware neuromorphic computing. In IEEE International Midwest Symposium on Circuits and Systems (MWSCAS).
[85]
Shihao Song, Anup Das, and Nagarajan Kandasamy. 2020. Improving dependability of neuromorphic computing with non-volatile memory. In European Dependable Computing Conference (EDCC).
[86]
Shihao Song, Lakshmi Varshika Mirtinti, Anup Das, and Nagarajan Kandasamy. 2021. A Design Flow for Mapping Spiking Neural Networks to Many-Core Neuromorphic Hardware. In International Conference on Computer-Aided Design (ICCAD).
[87]
Shihao Song, Twisha Titirsha, and Anup Das. 2021. Improving Inference Lifetime of Neuromorphic Systems via Intelligent Synapse Mapping. In International Conference on Application-specific Systems, Architectures and Processors (ASAP).
[88]
S. Sriram and S.S. Bhattacharyya. 2000. Embedded Multiprocessors; Scheduling and Synchronization. New York, NY, USA.
[89]
Sundararajan Sriram and Shuvra S Bhattacharyya. 2018. Embedded multiprocessors: Scheduling and synchronization. CRC press.
[90]
Stefan Steinke, Lars Wehmeyer, Bo-Sik Lee, and Peter Marwedel. 2002. Assigning program and data objects to scratchpad for energy reduction. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition. 409–415.
[91]
Sander Stuijk, Twan Basten, MCW Geilen, and Henk Corporaal. 2007. Multiprocessor resource allocation for throughput-constrained synchronous dataflow graphs. In Design Automation Conference. 777–782.
[92]
S. Stuijk, M. Geilen, and T. Basten. 2006. Exploring trade-offs in buffer requirements and throughput constraints for synchronous dataflow graphs. In Design Automation Conference (DAC).
[93]
Sander Stuijk, Marc Geilen, and Twan Basten. 2006. SDF 2303 3: SDF for free. In Sixth International Conference on Application of Concurrency to System Design (ACSD’06). IEEE, 276–278.
[94]
Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, and Alexander Alemi. 2017. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI conference on artificial intelligence, Vol.  31.
[95]
Amirhossein Tavanaei, Masoud Ghodrati, Saeed Reza Kheradpisheh, Timothée Masquelier, and Anthony Maida. 2019. Deep learning in spiking neural networks. Neural Networks 111(2019), 47–63.
[96]
Twisha Titirsha, Shihao Song, Anup Das, Jeffrey Krichmar, Nikil Dutt, Nagarajan Kandasamy, and Francky Catthoor. 2021. Endurance-aware mapping of spiking neural networks to neuromorphic hardware. IEEE Transactions on Parallel and Distributed Systems 33, 2 (2021), 288–301.
[97]
Haluk Topcuoglu, Salim Hariri, and Min-You Wu. 2002. Performance-effective and low-complexity task scheduling for heterogeneous computing. IEEE transactions on parallel and distributed systems 13, 3 (2002), 260–274.
[98]
Stavros Tripakis, Dai Bui, Marc Geilen, Bert Rodiers, and Edward A Lee. 2013. Compositionality in synchronous data flow: Modular code generation from hierarchical sdf graphs. ACM Transactions on Embedded Computing Systems (TECS) 12, 3(2013), 1–26.
[99]
ML Varshika, Abhishek Kumar Mishra, Nagarajan Kandasamy, and Anup Das. 2023. Hardware-Software Co-Design for On-Chip Learning in AI Systems. In Proceedings of the 28th Asia and South Pacific Design Automation Conference. 624–631.
[100]
M Lakshmi Varshika, Adarsha Balaji, Federico Corradi, Anup Das, Jan Stuijt, and Francky Catthoor. 2022. Design of Many-Core Big Little μBrains for Energy-Efficient Embedded Neuromorphic Computing. In Design, Automation, and Test in Europe (DATE) Conference and Exhibition.
[101]
M Lakshmi Varshika, Federico Corradi, and Anup Das. 2022. Nonvolatile Memories in Spiking Neural Network Architectures: Current and Emerging Trends. Electronics 11, 10 (2022), 1610.
[102]
Wayne Wolf, Ahmed Amine Jerraya, and Grant Martin. 2008. Multiprocessor system-on-chip (MPSoC) technology. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 27, 10(2008), 1701–1713.
[103]
Zhenxing Zhang, Shiyan Sun, Xunyu Chen, Tian Zhi, Qi Guo, and Yunji Chen. 2021. Space-address decoupled scratchpad memory management for neural network accelerators. Concurrency and Computation: Practice and Experience 33, 6(2021), e6046.

Cited By

View all
  • (2024)Neuromorphic Software Tools and Development EnvironmentsRevolutionizing AI with Brain-Inspired Technology10.4018/979-8-3693-6303-4.ch016(391-410)Online publication date: 4-Oct-2024
  • (2024)Neuromorphic Computing for the Masses2024 International Conference on Neuromorphic Systems (ICONS)10.1109/ICONS62911.2024.00014(39-46)Online publication date: 30-Jul-2024
  • (2024)An Open-Source And Extensible Framework for Fast Prototyping and Benchmarking of Spiking Neural Network Hardware2024 34th International Conference on Field-Programmable Logic and Applications (FPL)10.1109/FPL64840.2024.00042(250-256)Online publication date: 2-Sep-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Embedded Computing Systems
ACM Transactions on Embedded Computing Systems Just Accepted
EISSN:1558-3465
Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Journal Family

Publication History

Online AM: 02 December 2023
Accepted: 09 October 2023
Revised: 03 July 2023
Received: 31 March 2023

Check for updates

Author Tags

  1. spiking neural networks (SNNs)
  2. spiking deep convolutional neural networks (SDCNNs)
  3. neuromorphic hardware
  4. scratchpad memory
  5. interprocessor communication (IPC)
  6. integer linear programming (ILP)

Qualifiers

  • Research-article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)628
  • Downloads (Last 6 weeks)64
Reflects downloads up to 13 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Neuromorphic Software Tools and Development EnvironmentsRevolutionizing AI with Brain-Inspired Technology10.4018/979-8-3693-6303-4.ch016(391-410)Online publication date: 4-Oct-2024
  • (2024)Neuromorphic Computing for the Masses2024 International Conference on Neuromorphic Systems (ICONS)10.1109/ICONS62911.2024.00014(39-46)Online publication date: 30-Jul-2024
  • (2024)An Open-Source And Extensible Framework for Fast Prototyping and Benchmarking of Spiking Neural Network Hardware2024 34th International Conference on Field-Programmable Logic and Applications (FPL)10.1109/FPL64840.2024.00042(250-256)Online publication date: 2-Sep-2024
  • (2024)Towards Biology-Inspired Fault Tolerance of Neuromorphic Hardware for Space Applications2024 IEEE International Symposium on Defect and Fault Tolerance in VLSI and Nanotechnology Systems (DFT)10.1109/DFT63277.2024.10753531(1-7)Online publication date: 8-Oct-2024
  • (2023)Preserving Privacy of Neuromorphic Hardware From PCIe Congestion Side-Channel Attack2023 IEEE 47th Annual Computers, Software, and Applications Conference (COMPSAC)10.1109/COMPSAC57700.2023.00094(689-698)Online publication date: Jun-2023

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Full Access

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media