Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/3199700.3199802acmconferencesArticle/Chapter ViewAbstractPublication PagesiccadConference Proceedingsconference-collections
research-article

AEP: an error-bearing neural network accelerator for energy efficiency and model protection

Published: 13 November 2017 Publication History

Abstract

Neural Networks (NNs) have recently gained popularity in a wide range of modern application domains due to its superior inference accuracy. With growing problem size and complexity, modern NNs, e.g., CNNs (Convolutional NNs) and DNNs (Deep NNs), contain a large number of weights, which require tremendous efforts not only to prepare representative training datasets but also to train the network. There is an increasing demand to protect the NN weight matrices, an emerging Intellectual Property (IP) in NN field. Unfortunately, adopting conventional encryption method faces significant performance and energy consumption overheads.
In this paper, we propose AEP, a DianNao based NN accelerator design for IP protection. AEP aggressively reduces DRAM timing to generate a device dependent error mask, i.e., a set of erroneous cells while the distribution of these cells are device dependent due to process variations. AEP incorporates the error mask in the NN training process so that the trained weights are device dependent, which effectively defects IP piracy as exporting the weights to other devices cannot produce satisfactory inference accuracy. In addition, AEP speeds up NN inference and achieves significant energy reduction due to the fact that main memory dominates the energy consumption in DianNao accelerator. Our evaluation results show that by injecting 0.1% to 5% memory errors, AEP has negligible inference accuracy loss on the target device while exhibiting unacceptable accuracy degradation on other devices. In addition, AEP achieves an average of 72% performance improvement and 44% energy reduction over the DianNao baseline.

References

[1]
N. Ahmad, et al., "Low-power Compact Composite Field AES S-Box/Inv S-Box Design in 65nm CMOS using Novel XOR Gate," Integration, the VLSI Journal, 2013.
[2]
I. Bhati, et al., "DRAM Refresh Mechanisms, Penalties, And Trade-Offs," IEEE Transactions on Computers (TC), 2016.
[3]
T. Chen, et al., "DianNao: A Small-footprint High-throughput Accelerator For Ubiquitous Machine-learning," in ASPLOS, 2014.
[4]
Y.-H. Chen, et al., "Eyeriss: A Spatial Architecture For Energy-Efficient Dataflow For Convolutional Neural Networks," in ISCA, 2016.
[5]
P. Chi, et al., "PRIME: A Novel Processing-in-Memory Architecture For Neural Network Computation In ReRAM-Based Main Memory," in ISCA, 2016.
[6]
A. Coates, et al., "Deep learning with COTS HPC systems," in ICML, 2013.
[7]
M. Courbariaux, et al., "Training Deep Neural Networks With Low Precision Multiplications," arXiv:1409.1556, 2014.
[8]
M. Courbariaux, et al., "Binaryconnect: Training Deep Neural Networks with Binary Weights During Propagations," in NIPS, 2015.
[9]
N. Dowlin, et al., "Cryptonets: Applying Neural Networks to Encrypted Data with High Throughput and Accuracy," in ICML, 2016.
[10]
S. Gupta, et al., "Deep Learning with Limited Numerical Precision," in ICML, 2015.
[11]
G. Hinton, et al., "Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups," in IEEE Signal Processing Magazine, 2012.
[12]
ImageNet, "http://image-net.org/challenges/LSVRC/,".
[13]
JEDEC, "High Bandwidth Memory (HBM) DRAM," Technical report, JESD235A JC-42.3, 2015.
[14]
P. Judd, et al., "Reduced-precision Strategies for Bounded Memory In Deep Neural Nets," arXiv:1511.05236, 2015.
[15]
D. Kim, et al., "Neurocube: A Programmable Digital Neuromorphic Architecture With High-Density 3D Memory," in ISCA, 2016.
[16]
A. Krizhevsky, Learning Multiple Layers of Features from Tiny Images, Master's thesis, Univ. of Toronto, 2009.
[17]
Q. V. Le, "Building High-level Features Using Large Scale Unsupervised Learning," in ICASSP, 2013.
[18]
Y. LeCun, et al., "Gradient-Based Learning Applied to Document Recognition," Proc. of The IEEE, 1998.
[19]
D. Lee, et al., "Adaptive-latency DRAM: Optimizing DRAM Timing For The Common-case," in HPCA, 2015.
[20]
D. Lie, et al., "Architectural Support for Copy and Tamper Resistant Software," in ASPLOS, 2000.
[21]
S. Liu, et al., "Flikker: Saving DRAM Refresh-power through Critical Data Partitioning," in ASPLOS, 2012.
[22]
Y. Netzer, et al., "Reading Digits in Natural Images with Unsupervised Feature Learning," in NIPS, 2011.
[23]
K. Ni, et al., "Large-scale deep learning on the YFCC100M dataset," arXiv:1502.03409, 2015.
[24]
M. K. Qureshi, et al., "AVATAR: A Variable-Retention-Time (VRT) Aware Refresh For DRAM Systems," in DSN, 2015.
[25]
P. Rosenfeld, et al., "DRAMSim2: A Cycle Accurate Memory System Simulator," in IEEE Computer Architecture Letters, 2011.
[26]
A. Shafiee, et al., "ISAAC: A Convolutional Neural Network Accelerator With In-Situ Analog Arithmetic In Crossbars," in ISCA, 2016.
[27]
K. Simonyan, et al., "Very Deep Convolutional Networks for Large-scale Image Recognition," arXiv:1409.1556, 2014.
[28]
M. Song, et al., "Towards Pervasive And User Satisfactory CNN Across GPU Microarchitectures," in HPCA, 2017.
[29]
G. E. Suh, et al., "Physical Unclonable Functions for Device Authentication and Secret Key Generation," in DAC, 2007.
[30]
TheanoTeam, "Theano: A Python Framework for Fast Computation of Mathematical Expressions," arXiv:1605.02688, 2016.
[31]
X. Zhang, et al., "Exploiting DRAM Restore Time Variations In Deep Sub-micron Scaling," in DATE, 2015.
[32]
X. Zhang, et al., "Restore Truncation For Performance Improvement In Future DRAM Systems," in HPCA, 2016.
[33]
Y. Zhang, et al., "Architectural Support for Protecting User Privacy on Trusted Processors," in ASPLOS, 2004.

Cited By

View all
  1. AEP: an error-bearing neural network accelerator for energy efficiency and model protection

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    ICCAD '17: Proceedings of the 36th International Conference on Computer-Aided Design
    November 2017
    1077 pages

    Sponsors

    In-Cooperation

    • IEEE-EDS: Electronic Devices Society

    Publisher

    IEEE Press

    Publication History

    Published: 13 November 2017

    Check for updates

    Qualifiers

    • Research-article

    Conference

    ICCAD '17
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 457 of 1,762 submissions, 26%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)2
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 04 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media