Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3349567.3351726acmotherconferencesArticle/Chapter ViewAbstractPublication PagesesweekConference Proceedingsconference-collections
research-article

DORY: Lightweight memory hierarchy management for deep NN inference on IoT endnodes: work-in-progress

Published: 13 October 2019 Publication History

Abstract

IoT endnodes often couple a small and fast L1 scratchpad memory with higher-capacity but lower bandwidth and speed L2 background memory. The absence of a coherent hardware cache hierarchy saves energy but comes at the cost of labor-intensive explicit memory management, complicating the deployment of algorithms with large data memory footprint, such as Deep Neural Network (DNN) inference. In this work, we present DORY, a lightweight software-cache dedicated to DNN Deployment Oriented to memoRY. DORY leverages static data tiling and DMA-based double buffering to hide the complexity of manual L1-L2 memory traffic management. DORY enables storage of activations and weights in L2 with less than 4% performance overhead with respect to direct execution in L1. We show that a 142 kB DNN achieving 79.9% on CIFAR-10 runs 3.2X faster compared to its execution directly from L2 memory while consuming 1.9X less energy.

References

[1]
Google AI. 2015. OR tools. (2015). https://developers.google.com/optimization/
[2]
Leonardo Cecconi et al., 2017. Optimal tiling strategy for memory bandwidth reduction for cnns. In International Conference on ACIVS. Springer, 89--100.
[3]
E. Flamand et al, 2018. GAP-8: A RISC-V SoC for AI at the Edge of the IoT. In 2018 IEEE 29th ASAP International Conference. 1--4.
[4]
Daniele Palossi et al, 2019. A 64mW DNN-based Visual Navigation Engine for Autonomous Nano-Drones. IEEE Internet of Things Journal (2019).
[5]
Maurice Peemen et al., 2013. Memory-centric accelerator design for convolutional neural networks. In 2013 IEEE 31st ICCD.
[6]
Christian Pinto et al., 2013. A highly efficient, thread-safe software cache implementation for tightly-coupled multicore clusters. In IEEE 24th ASAP Conference
[7]
Selma Saidi et al., 2013. Optimizing two-dimensional DMA transfers for scratch-pad Based MPSoCs platforms. Microprocessors and Microsystems.

Cited By

View all
  • (2024)TinyForge: A Design Space Exploration to Advance Energy and Silicon Area Trade-offs in tinyML Compute Architectures with Custom Latch ArraysProceedings of the 29th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 310.1145/3620666.3651328(1033-1047)Online publication date: 27-Apr-2024
  • (2021)A Fully Integrated 5-mW, 0.8-Gbps Energy-Efficient Chip-to-Chip Data Link for Ultralow-Power IoT End-Nodes in 65-nm CMOSIEEE Transactions on Very Large Scale Integration (VLSI) Systems10.1109/TVLSI.2021.310880629:10(1800-1811)Online publication date: Oct-2021
  • (2021)Energy-Efficient Hardware-Accelerated Synchronization for Shared-L1-Memory Multiprocessor ClustersIEEE Transactions on Parallel and Distributed Systems10.1109/TPDS.2020.302869132:3(633-648)Online publication date: 1-Mar-2021
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
CODES/ISSS '19: Proceedings of the International Conference on Hardware/Software Codesign and System Synthesis Companion
October 2019
64 pages
ISBN:9781450369237
DOI:10.1145/3349567
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 13 October 2019

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article

Conference

CODES/ISSS '19

Acceptance Rates

Overall Acceptance Rate 280 of 864 submissions, 32%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)34
  • Downloads (Last 6 weeks)2
Reflects downloads up to 08 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)TinyForge: A Design Space Exploration to Advance Energy and Silicon Area Trade-offs in tinyML Compute Architectures with Custom Latch ArraysProceedings of the 29th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 310.1145/3620666.3651328(1033-1047)Online publication date: 27-Apr-2024
  • (2021)A Fully Integrated 5-mW, 0.8-Gbps Energy-Efficient Chip-to-Chip Data Link for Ultralow-Power IoT End-Nodes in 65-nm CMOSIEEE Transactions on Very Large Scale Integration (VLSI) Systems10.1109/TVLSI.2021.310880629:10(1800-1811)Online publication date: Oct-2021
  • (2021)Energy-Efficient Hardware-Accelerated Synchronization for Shared-L1-Memory Multiprocessor ClustersIEEE Transactions on Parallel and Distributed Systems10.1109/TPDS.2020.302869132:3(633-648)Online publication date: 1-Mar-2021
  • (2021)ZigZag: Enlarging Joint Architecture-Mapping Design Space Exploration for DNN AcceleratorsIEEE Transactions on Computers10.1109/TC.2021.305996270:8(1160-1174)Online publication date: 1-Aug-2021
  • (2021)Predicting Hard Disk Failures in Data Centers Using Temporal Convolutional Neural NetworksEuro-Par 2020: Parallel Processing Workshops10.1007/978-3-030-71593-9_22(277-289)Online publication date: 14-Mar-2021
  • (2020)Memory-Latency-Accuracy Trade-Offs for Continual Learning on a RISC-V Extreme-Edge Node2020 IEEE Workshop on Signal Processing Systems (SiPS)10.1109/SiPS50750.2020.9195220(1-6)Online publication date: Oct-2020
  • (2020)Q-EEGNet: an Energy-Efficient 8-bit Quantized Parallel EEGNet Implementation for Edge Motor-Imagery Brain-Machine Interfaces2020 IEEE International Conference on Smart Computing (SMARTCOMP)10.1109/SMARTCOMP50058.2020.00065(284-289)Online publication date: Sep-2020

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media