Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3489517.3530592acmconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article

SATO: spiking neural network acceleration via temporal-oriented dataflow and architecture

Published: 23 August 2022 Publication History
  • Get Citation Alerts
  • Abstract

    Event-driven spiking neural networks (SNNs) have shown great promise for being strikingly energy-efficient. SNN neurons integrate the spikes, accumulate the membrane potential, and fire output spike when the potential exceeds a threshold. Existing SNN accelerators, however, have to carry out such accumulation-comparison operation in serial. Repetitive spike generation at each time step not only increases latency as well as overall energy budget, but also incurs memory access overhead of fetching membrane potentials, both of which lessen the efficiency of SNN accelerators. Meanwhile, inherent highly sparse spikes of SNNs lead to imbalanced workloads among neurons that hurdle the utilization of processing elements (PEs).
    This paper proposes SATO, a temporal-parallel SNN accelerator that accumulates the membrane potential for all time steps in parallel. SATO architecture contains a novel binary adder-search tree to generate the output spike train, which decouples the chronological dependence in the accumulation-comparison operation. Moreover, SATO can evenly dispatch the compressed workloads to all PEs with maximized data locality of input spike trains based on a bucket-sort-based method. Our evaluations show that SATO outperforms the previous ANN accelerator 8-bit version of "Eyeriss" by 30.9× in terms of speedup and 12.3×, in terms of energy-saving. Compared with the state-of-the-art SNN accelerator "SpinalFlow", SATO can also achieve 6.4× performance gain and 4.8× energy reduction, which is quite impressive for inference.

    References

    [1]
    Filipp Akopyan et al. 2015. Truenorth: Design and tool flow of a 65 mw 1 million neuron programmable neurosynaptic chip. TCAD (2015).
    [2]
    Ben Varkey Benjamin et al. 2014. Neurogrid: A mixed-analog-digital multichip system for large-scale neural simulations. Proc. of the IEEE (2014).
    [3]
    Yu-Hsin Chen et al. 2016. Eyeriss: A spatial architecture for energy-efficient dataflow for CNNs. ACM SIGARCH Computer Architecture News (2016).
    [4]
    Iulia M Comsa et al. 2020. Temporal coding in spiking neural networks with alpha synaptic function. In ICASSP. IEEE.
    [5]
    Mike Davies et al. 2018. Loihi: A neuromorphic manycore processor with on-chip learning. Ieee Micro (2018).
    [6]
    Lei Deng et al. 2020. Rethinking the performance comparison between SNNS and ANNS. Neural Networks (2020).
    [7]
    Wei Fang et al. 2021. Incorporating learnable membrane time constant to enhance learning of spiking neural networks. In ICCV.
    [8]
    Bing Han and Kaushik Roy. 2021. Deep Spiking Neural Network: Energy Efficiency Through Time based Coding. In ECCV. 388--404.
    [9]
    Alireza Khodamoradi, Kristof Denolf, and Ryan Kastner. 2021. S2N2: A FPGA Accelerator for Streaming Spiking Neural Networks. In FPGA. 194--205.
    [10]
    Yang Li et al. 2021. BSNN: Towards Faster and Better Conversion of ANNs to SNNs with Bistable Neurons. arXiv (2021).
    [11]
    Fangxin Liu et al. 2020. SSTDP: Supervised Spike Timing Dependent Plasticity for Efficient Spiking Neural Network Training. Frontiers in Neuroscience (2020).
    [12]
    Maryam Mirsadeghi et al. 2021. STiDi-BP: Spike time displacement based error backpropagation in multilayer SNNs. Neurocomputing (2021).
    [13]
    Naveen Muralimanohar et al. 2009. CACTI 6.0: A tool to understand large caches. University of Utah and Hewlett Packard Laboratories, Tech. Rep (2009).
    [14]
    Surya Narayanan et al. 2020. SpinalFlow: an architecture and dataflow tailored for spiking neural networks. In ISCA. IEEE.
    [15]
    Seongsik Park and Sungroh Yoon. 2021. Training Energy-Efficient Deep Spiking Neural Networks with Time-to-First-Spike Coding. arXiv (2021).
    [16]
    Kaushik Roy et al. 2019. Towards spike-based machine intelligence with neuromorphic computing. Nature (2019).
    [17]
    Abhronil Sengupta et al. 2019. Going deeper in spiking neural networks: VGG and residual architectures. Frontiers in neuroscience (2019).
    [18]
    Sonali Singh et al. 2020. NEBULA: a neuromorphic spin-based ultra-low power architecture for SNNs and ANNs. In ISCA. IEEE.
    [19]
    Synopsys. [Online]. https://www.synopsys.com/community/university-program/teaching-resources.html.
    [20]
    Amirhossein Tavanaei and Anthony Maida. 2019. BP-STDP: Approximating backpropagation using spike timing dependent plasticity. Neurocomputing (2019).
    [21]
    Yaman Umuroglu et al. 2018. Bismo: A scalable bit-serial matrix multiplication overlay for reconfigurable computing. In FPL. IEEE.

    Cited By

    View all
    • (2024)EPHA: An Energy-efficient Parallel Hybrid Architecture for ANNs and SNNsACM Transactions on Design Automation of Electronic Systems10.1145/364313429:3(1-28)Online publication date: 14-Mar-2024
    • (2024)SPAT: FPGA-based Sparsity-Optimized Spiking Neural Network Training Accelerator with Temporal Parallel Dataflow2024 IEEE International Symposium on Circuits and Systems (ISCAS)10.1109/ISCAS58744.2024.10558351(1-5)Online publication date: 19-May-2024
    • (2024)Stellar: Energy-Efficient and Low-Latency SNN Algorithm and Hardware Co-Design with Spatiotemporal Computation2024 IEEE International Symposium on High-Performance Computer Architecture (HPCA)10.1109/HPCA57654.2024.00023(172-185)Online publication date: 2-Mar-2024
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    DAC '22: Proceedings of the 59th ACM/IEEE Design Automation Conference
    July 2022
    1462 pages
    ISBN:9781450391429
    DOI:10.1145/3489517
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 23 August 2022

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    DAC '22
    Sponsor:
    DAC '22: 59th ACM/IEEE Design Automation Conference
    July 10 - 14, 2022
    California, San Francisco

    Acceptance Rates

    Overall Acceptance Rate 1,770 of 5,499 submissions, 32%

    Upcoming Conference

    DAC '25
    62nd ACM/IEEE Design Automation Conference
    June 22 - 26, 2025
    San Francisco , CA , USA

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)182
    • Downloads (Last 6 weeks)27
    Reflects downloads up to 27 Jul 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)EPHA: An Energy-efficient Parallel Hybrid Architecture for ANNs and SNNsACM Transactions on Design Automation of Electronic Systems10.1145/364313429:3(1-28)Online publication date: 14-Mar-2024
    • (2024)SPAT: FPGA-based Sparsity-Optimized Spiking Neural Network Training Accelerator with Temporal Parallel Dataflow2024 IEEE International Symposium on Circuits and Systems (ISCAS)10.1109/ISCAS58744.2024.10558351(1-5)Online publication date: 19-May-2024
    • (2024)Stellar: Energy-Efficient and Low-Latency SNN Algorithm and Hardware Co-Design with Spatiotemporal Computation2024 IEEE International Symposium on High-Performance Computer Architecture (HPCA)10.1109/HPCA57654.2024.00023(172-185)Online publication date: 2-Mar-2024
    • (2023)FireFly: A High-Throughput Hardware Accelerator for Spiking Neural Networks With Efficient DSP and Memory OptimizationIEEE Transactions on Very Large Scale Integration (VLSI) Systems10.1109/TVLSI.2023.327934931:8(1178-1191)Online publication date: 1-Aug-2023
    • (2022)Randomize and Match: Exploiting Irregular Sparsity for Energy Efficient Processing in SNNs2022 IEEE 40th International Conference on Computer Design (ICCD)10.1109/ICCD56317.2022.00073(451-454)Online publication date: Oct-2022

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media