Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3349567.3351730acmotherconferencesArticle/Chapter ViewAbstractPublication PagesesweekConference Proceedingsconference-collections
research-article
Public Access

Offloading cache configuration prediction to an FPGA for hardware speedup and overhead reduction: work-in-progress

Published: 13 October 2019 Publication History

Abstract

In this paper, we present our cache configuration prediction methodology offloaded to an FPGA for improved performance and hardware overhead reduction, while maintaining cache configuration predictions within 5% of the optimal energy cache configuration for application phases for the instruction and data caches.

References

[1]
M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, M. Kudlur, J. Levenberg, R. Monga, S. Moore, D. G. Murray, B. Steiner, P. Tucker, V. Vasudevan, P. Warden, M. Wicke, Y. Yu and X. Zheng, "TensorFlow: A system for large-scale machine learning," in 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), 2016
[2]
arXiv:1807.05317 [cs.LG].
[3]
B. Dutta, V. Adhinarayanan, and W. Feng. 2018. GPU power prediction via ensemble machine learning for DVFS space exploration. In Proceedings of the 15th ACM International Conference on Computing Frontiers (CF '18). ACM, New York, NY, USA, 240--243.
[4]
EEMBC. The Embedded Microprocessor Benchmark Consortium http://www.eembc.org/benchmark/automotive_sl.php, Sept. 2013
[5]
V. Gokhale, J. Jin, A. Dundar, B. Martini and E. Culurciello, "A 240 G-ops/s Mobile Coprocessor for Deep Neural Networks," 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, 2014, pp. 696--701.
[6]
A. Gordon-Ross, F. Vahid and N. Dutt, "Fast configurable-cache tuning with a unified second-level cache," ISLPED '05. Proceedings of the 2005 International Symposium on Low Power Electronics and Design, 2005., San Diego, CA, 2005, pp. 323--326.
[7]
M. R. Guthaus, J. S. Ringenberg, D. Ernst, T. M. Austin, T. Mudge, and R. B. Brown. 2001. MiBench: A free, commercially representative embedded benchmark suite. In Proceedings of the Workload Characterization, 2001. WWC-4. 2001 IEEE International Workshop (WWC '01). IEEE Computer Society, Washington, DC, USA, 3--14.
[8]
S. Khakhaeng and C. Chantrapornchai, "On the finding proper cache prediction model using neural network," 2016 8th International Conference on Knowledge and Smart Technology (KST), Chiangmai, 2016, pp. 146--151.
[9]
N. D. Lane et al., "DeepX: A Software Accelerator for Low-Power Deep Learning Inference on Mobile Devices," 2016 15th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN), Vienna, 2016, pp. 1--12.
[10]
"Perf events tutorial," http://perf.wiki.kernel.org/, 2012.

Cited By

View all
  • (2023)Artificial Intelligence empowered content caching for energy optimization in vehicular networksProceedings of the 5th International Conference on Information Management & Machine Intelligence10.1145/3647444.3647847(1-4)Online publication date: 23-Nov-2023
  • (2022)Collaborative Caching for Energy Optimization in Content-Centric Internet of ThingsIEEE Transactions on Computational Social Systems10.1109/TCSS.2021.30871979:1(230-238)Online publication date: Feb-2022
  1. Offloading cache configuration prediction to an FPGA for hardware speedup and overhead reduction: work-in-progress

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    CODES/ISSS '19: Proceedings of the International Conference on Hardware/Software Codesign and System Synthesis Companion
    October 2019
    64 pages
    ISBN:9781450369237
    DOI:10.1145/3349567
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 October 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. FPGA
    2. artificial neural network
    3. cache
    4. configurable
    5. machine learning
    6. mobile devices
    7. phases

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    CODES/ISSS '19

    Acceptance Rates

    Overall Acceptance Rate 280 of 864 submissions, 32%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)39
    • Downloads (Last 6 weeks)7
    Reflects downloads up to 15 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2023)Artificial Intelligence empowered content caching for energy optimization in vehicular networksProceedings of the 5th International Conference on Information Management & Machine Intelligence10.1145/3647444.3647847(1-4)Online publication date: 23-Nov-2023
    • (2022)Collaborative Caching for Energy Optimization in Content-Centric Internet of ThingsIEEE Transactions on Computational Social Systems10.1109/TCSS.2021.30871979:1(230-238)Online publication date: Feb-2022

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media