Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Adversarial Coreset Selection for Efficient Robust Training

Published: 14 August 2023 Publication History

Abstract

It has been shown that neural networks are vulnerable to adversarial attacks: adding well-crafted, imperceptible perturbations to their input can modify their output. Adversarial training is one of the most effective approaches to training robust models against such attacks. Unfortunately, this method is much slower than vanilla training of neural networks since it needs to construct adversarial examples for the entire training data at every iteration. By leveraging the theory of coreset selection, we show how selecting a small subset of training data provides a principled approach to reducing the time complexity of robust training. To this end, we first provide convergence guarantees for adversarial coreset selection. In particular, we show that the convergence bound is directly related to how well our coresets can approximate the gradient computed over the entire training data. Motivated by our theoretical analysis, we propose using this gradient approximation error as our adversarial coreset selection objective to reduce the training set size effectively. Once built, we run adversarial training over this subset of the training data. Unlike existing methods, our approach can be adapted to a wide variety of training objectives, including TRADES, p-PGD, and Perceptual Adversarial Training. We conduct extensive experiments to demonstrate that our approach speeds up adversarial training by 2–3 times while experiencing a slight degradation in the clean and robust accuracy.

References

[1]
Adadi A A survey on data-efficient algorithms in big data era Journal of Big Data 2021 8 1 1-54
[2]
Andriushchenko, M., & Flammarion, N. (2020). Understanding and improving fast adversarial training. In Proceedings of the advances in neural information processing systems 33: Annual conference on neural information processing systems.
[3]
Biggio, B., Corona, I., & Maiorca, D., et al. (2013). Evasion attacks against machine learning at test time. In Proceedings of the European conference on machine learning and knowledge discovery in databases (ECML-PKDD), pp. 387–402.
[4]
Campbell, T., & Broderick, T. (2018). Bayesian coreset construction via greedy iterative geodesic ascent. In Proceedings of the 35th international conference on machine learning (ICML), pp. 697–705.
[5]
Croce, F., & Hein, M. (2020). Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In Proceedings of the 37th international conference on machine learning (ICML) 2020, pp. 2206–2216.
[6]
Danskin JM The theory of max-min and its application to weapons allocation problems 1967 Springer Science & Business Media
[7]
Elenberg, E. R., Khanna, R., & Dimakis, A. G., et al. (2016). Restricted strong convexity implies weak submodularity. CoRR abs/1612.00804, arXiv:1612.00804.
[8]
Eykholt, K., Evtimov, I., & Fernandes, E., et al. (2018). Robust physical-world attacks on deep learning visual classification. In Proceeding of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 1625–1634.
[9]
Feldman, D. (2020). Introduction to core-sets: An updated survey. CoRR abs/2011.09384. arXiv:2011.09384.
[10]
Goodfellow, I. J., Shlens, J., & Szegedy, C. (2015). Explaining and harnessing adversarial examples. In Proceedings of the 3rd international conference on learning representations (ICLR).
[11]
Har-Peled, S., & Mazumdar, S. (2004). On coresets for k-means and k-median clustering. In Proceedings of the 36th annual ACM symposium on theory of computing (STOC), pp. 291–300.
[12]
He, K., Zhang, X., & Ren, S., et al. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 770–778.
[13]
de Jorge Aranda, P., Bibi, A., & Volpi, R., et al. (2022). Make some noise: Reliable and efficient single-step adversarial training. In Proceedings of the advances in neural information processing systems 35: Annual conference on neural information processing systems (NeurIPS).
[14]
Kang, D., Sun, Y., & Hendrycks, D., et al. (2019). Testing robustness against unforeseen adversaries. CoRR abs/1908.08016, arXiv:1908.08016.
[15]
Karras, T., Laine, S., & Aittala, M., et al. (2020). Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp 8107–8116.
[16]
Katharopoulos, A., & Fleuret, F. (2018). Not all samples are created equal: Deep learning with importance sampling. In Proceedings of the 35th international conference on machine learning (ICML), pp 2530–2539.
[17]
Killamsetty, K., Sivasubramanian, D., & Ramakrishnan, G., et al. (2021a). GRAD-MATCH: gradient matching based data subset selection for efficient deep model training. In Proceedings of the 38th international conference on machine learning (ICML), pp. 5464–5474.
[18]
Killamsetty, K., Sivasubramanian, D., & Ramakrishnan, G., et al. (2021b). GLISTER: Generalization based data subset selection for efficient and robust learning. In Proceedings of the 35th AAAI conference on artificial intelligence, pp. 8110–8118.
[19]
Killamsetty, K., Zhao, X., & Chen, F., et al. (2021c). RETRIEVE: Coreset selection for efficient and robust semi-supervised learning. In Advances in neural information processing systems 34: Annual conference on neural information processing systems (NeurIPS), pp. 14488–14501.
[20]
Kolter, Z., & Madry, A. (2018). Adversarial robustness: Theory and practice. In Tutorial in the advances in neural information processing systems 31: Annual conference on neural information processing systems (NeurIPS). https://adversarial-ml-tutorial.org/.
[21]
Krizhevsky, A., & Hinton, G. (2009). Learning multiple layers of features from tiny images. Master’s thesis, Department of Computer Science, University of Toronto.
[22]
Laidlaw, C., & Feizi, S. (2019). Functional adversarial attacks. In Proceedings of the advances in neural information processing systems 32: Annual conference on neural information processing systems (NeurIPS), pp. 10408–10418.
[23]
Laidlaw, C., Singla, S., & Feizi, S. (2021). Perceptual adversarial robustness: Defense against unseen threat models. In Proceedings of the 9th international conference on learning representations (ICLR).
[24]
Liu, Y., Ma, X., & Bailey, J., et al. (2020). Reflection backdoor: A natural backdoor attack on deep neural networks. In Proceedings of the 16th European conference on computer vision (ECCV), pp. 182–199.
[25]
Ma X, Niu Y, Gu L, et al. Understanding adversarial attacks on deep learning based medical image analysis systems Pattern Recognition 2021 110 107 332
[26]
Madry, A., Makelov, A., & Schmidt, L., et al. (2018). Towards deep learning models resistant to adversarial attacks. In Proceedings of the 6th international conference on learning representations (ICLR).
[27]
Minoux, M. (1978). Accelerated greedy algorithms for maximizing submodular set functions. In Optimization techniques (pp. 234–243). Springer.
[28]
Mirzasoleiman, B., Bilmes, J. A., & Leskovec, J. (2020a). Coresets for data-efficient training of machine learning models. In Proceedings of the 37th international conference on machine learning (ICML), pp. 6950–6960.
[29]
Mirzasoleiman, B., Cao, K., & Leskovec, J. (2020b). Coresets for robust training of deep neural networks against noisy labels. In Proceedings of the advances in neural information processing systems 33: annual conference on neural information processing systems (NeurIPS).
[30]
Nemhauser GL, Wolsey LA, and Fisher ML An analysis of approximations for maximizing submodular set functions: I Mathematical Programming 1978 14 1 265-294
[31]
Netzer, Y., Wang, T., & Coates, A., et al. (2011). Reading digits in natural images with unsupervised feature learning. In NeurIPS workshop on deep learning and unsupervised feature learning.
[32]
Pati, Y. C., Rezaiifar, R., & Krishnaprasad, P. S. (1993). Orthogonal matching pursuit: recursive function approximation with applications to wavelet decomposition. In Proceedings of 27th asilomar conference on signals, systems and computers, pp. 40–44.
[33]
Qin, C., Martens, J., & Gowal, S., et al. (2019). Adversarial robustness through local linearization. In Proceedings of the advances in neural information processing systems 32: Annual conference on neural information processing systems (NeurIPS).
[34]
Russakovsky O, Deng J, Su H, et al. ImageNet large scale visual recognition challenge International Journal of Computer Vision 2015 115 3 211-252
[35]
Schwartz, R., Dodge, J., Smith, N. A., et al. (2020). Green AI. Communication of the ACM,63(12), 54–63.
[36]
Smith, L. N. (2017). Cyclical learning rates for training neural networks. In Proceedings of the IEEE winter conference on applications of computer vision (WACV), pp. 464–472
[37]
Strubell, E., Ganesh, A., & McCallum, A. (2019). Energy and policy considerations for deep learning in NLP. In Proceedings of the 57th conference of the association for computational linguistics ACL, pp. 3645–3650.
[38]
Szegedy, C., Zaremba, W., & Sutskever, I., et al. (2014). Intriguing properties of neural networks. In Proceedings of the 2nd international conference on learning representations (ICLR).
[39]
Tramèr, F., Kurakin, A., & Papernot, N., et al. (2018). Ensemble adversarial training: Attacks and defenses. In Proceedings of the 6th international conference on learning representations (ICLR).
[40]
Tsipras, D., Santurkar, S., & Engstrom, L., et al. (2019). Robustness may be at odds with accuracy. In Proceedings of the 7th international conference on learning representations (ICLR).
[41]
Vahdat, A., & Kautz, J. (2020). NVAE: A deep hierarchical variational autoencoder. In Proceedings of the advances in neural information processing systems 33: Annual conference on neural information processing systems (NeurIPS).
[42]
Wei, K., Iyer, R., & Bilmes, J. (2015). Submodularity in data subset selection and active learning. In Proceedings of the 32nd international conference on machine learning (ICML), pp. 1954–1963.
[43]
Wolsey LA An analysis of the greedy algorithm for the submodular set covering problem Combinatorica 1982 2 4 385-393
[44]
Wong, E., Rice, L., & Kolter, J. Z. (2020). Fast is better than free: Revisiting adversarial training. In Proceedings of the 8th international conference on learning representations (ICLR).
[45]
Wu, Y., Kirillov, A., & Massa, F., et al. (2019). Detectron2. https://github.com/facebookresearch/detectron2.
[46]
Xiao, C., Zhu, J., & Li, B., et al. (2018). Spatially transformed adversarial examples. In Proceedings of the 6th international conference on learning representations (ICLR).
[47]
Zagoruyko, S., & Komodakis, N. (2016). Wide residual networks. In Proceedings of the British machine vision conference (BMVC).
[48]
Zhang, H., Yu, Y., & Jiao, J., et al. (2019). Theoretically principled trade-off between robustness and accuracy. In Proceedings of the 36th international conference on machine learning (ICML), pp. 7472–7482
[49]
Zhang, R., Isola, P., & Efros, A. A., et al. (2018). The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 586–595

Cited By

View all
  • (2024)Artificial Immune System of Secure Face Recognition Against Adversarial AttacksInternational Journal of Computer Vision10.1007/s11263-024-02153-0132:12(5718-5740)Online publication date: 1-Dec-2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image International Journal of Computer Vision
International Journal of Computer Vision  Volume 131, Issue 12
Dec 2023
232 pages

Publisher

Kluwer Academic Publishers

United States

Publication History

Published: 14 August 2023
Accepted: 19 July 2023
Received: 02 January 2023

Author Tags

  1. Adversarial training
  2. Coreset selection
  3. Efficient training
  4. Robust deep learning
  5. Image classification

Qualifiers

  • Research-article

Funding Sources

  • University of Melbourne

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 20 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Artificial Immune System of Secure Face Recognition Against Adversarial AttacksInternational Journal of Computer Vision10.1007/s11263-024-02153-0132:12(5718-5740)Online publication date: 1-Dec-2024

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media