Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3289602.3293947acmconferencesArticle/Chapter ViewAbstractPublication PagesfpgaConference Proceedingsconference-collections
poster

Overcoming Data Transfer Bottlenecks in DNN Accelerators via Layer-Conscious Memory Managment

Published: 20 February 2019 Publication History

Abstract

Deep Neural Networks (DNNs) are rapidly evolving to satisfy the performance and accuracy requirements in many real world applications. The evolution renders DNNs more and more complex in terms of network topology, data sizes and layer types. Currently most state-of-the-art DNN accelerators adopt a uniform memory hierarchy (UMH) design methodology, which means that the data transferring of all convolutional and fully connected layers must go through the same memory levels. Unfortunately, for some layers, the performance is always bounded by off-chip memory transferring. It is caused by the saturating of data reuse happening in on-chip buffers, resulting in underutilization of on-chip memory. To address this issue, we propose a layer-conscious memory hierarchy (LCMH) methodology for DNN accelerators. LCMH could determine the memory levels of all the layers according to their requirements for off-chip memory bandwidth and on-chip buffer size for the data sources. As a result, the off-chip memory footprints of memory bounded layers could be avoided by keeping the data of them on chip. In addition, we provide architectural support for the accelerators equipped with LCMH. Experimental results show that designs with layer- conscious memory management could achieve up to 36% speedup compared with the designs wth UMH and 5% improvement over state-of-the-art designs.

Cited By

View all
  • (2024)A Survey on Neural Network Hardware AcceleratorsIEEE Transactions on Artificial Intelligence10.1109/TAI.2024.33771475:8(3801-3822)Online publication date: Aug-2024
  • (2022)Memory-aware Partitioning, Scheduling, and Floorplanning for Partially Dynamically Reconfigurable SystemsACM Transactions on Design Automation of Electronic Systems10.1145/353496828:1(1-21)Online publication date: 23-May-2022
  • (2022)Non-Structured DNN Weight Pruning—Is It Beneficial in Any Platform?IEEE Transactions on Neural Networks and Learning Systems10.1109/TNNLS.2021.306326533:9(4930-4944)Online publication date: Sep-2022
  • Show More Cited By

Index Terms

  1. Overcoming Data Transfer Bottlenecks in DNN Accelerators via Layer-Conscious Memory Managment

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    FPGA '19: Proceedings of the 2019 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays
    February 2019
    360 pages
    ISBN:9781450361378
    DOI:10.1145/3289602
    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 20 February 2019

    Check for updates

    Author Tags

    1. accelerator
    2. dnns
    3. fpga
    4. memory management

    Qualifiers

    • Poster

    Funding Sources

    • Beijing Natural Science Foundation
    • Falcon Computing Solutions Inc.

    Conference

    FPGA '19
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 125 of 627 submissions, 20%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 06 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)A Survey on Neural Network Hardware AcceleratorsIEEE Transactions on Artificial Intelligence10.1109/TAI.2024.33771475:8(3801-3822)Online publication date: Aug-2024
    • (2022)Memory-aware Partitioning, Scheduling, and Floorplanning for Partially Dynamically Reconfigurable SystemsACM Transactions on Design Automation of Electronic Systems10.1145/353496828:1(1-21)Online publication date: 23-May-2022
    • (2022)Non-Structured DNN Weight Pruning—Is It Beneficial in Any Platform?IEEE Transactions on Neural Networks and Learning Systems10.1109/TNNLS.2021.306326533:9(4930-4944)Online publication date: Sep-2022
    • (2022)FCNNLib: A Flexible Convolution Algorithm Library for Deep Learning on FPGAsIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2021.310806541:8(2546-2559)Online publication date: Aug-2022
    • (2022)Polymorphic Accelerators for Deep Neural NetworksIEEE Transactions on Computers10.1109/TC.2020.304862471:3(534-546)Online publication date: 1-Mar-2022
    • (2022)Tetris: A Heuristic Static Memory Management Framework for Uniform Memory Multicore Neural Network AcceleratorsJournal of Computer Science and Technology10.1007/s11390-021-1213-337:6(1255-1270)Online publication date: 30-Nov-2022
    • (2021)OMNI: A Framework for Integrating Hardware and Software Optimizations for Sparse CNNsIEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems10.1109/TCAD.2020.302390340:8(1648-1661)Online publication date: Aug-2021
    • (2020)Learning in the Frequency Domain2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)10.1109/CVPR42600.2020.00181(1737-1746)Online publication date: Jun-2020
    • (2020)Survey on memory management techniques in heterogeneous computing systemsIET Computers & Digital Techniques10.1049/iet-cdt.2019.009214:2(47-60)Online publication date: 21-Jan-2020

    View Options

    Get Access

    Login options

    View options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media