271 Peeyush
271 Peeyush
271 Peeyush
in
International Conference on Emerging applications of Artificial Intelligence,
Machine Learning and Cybersecurity (ICAMC-2024)
At
HMRITM, New Delhi, India.
On
May 16-17, 2024
01 02 03
Security Implications Defense Strategies
Vulnerability of ML
Various defense strategies have been
Models These adversarial examples, often
proposed, with denoising autoencoders
imperceptible to humans, pose a significant
showing promise in filtering out
Machine learning models, especially deep security threat in critical applications such
adversarial perturbations by learning to
neural networks, are vulnerable to as autonomous driving and medical
reconstruct original inputs from noisy
adversarial attacks, which are small, diagnosis.
versions.
carefully crafted perturbations in input data
that lead to incorrect model outputs.
Autoencoder Architecture
Denoising Autoencoders consist of an encoder that reduces input data to a lower-
dimensional representation and a decoder that attempts to reconstruct the
FGSM Attack
The Fast Gradient Sign Method (FGSM) is a common adversarial technique that
creates perturbations by manipulating the gradient of the loss function to mislead
the model.
The Power of Ensemble Learning
Basics of Ensemble Learning:
•Ensemble learning combines predictions from multiple models.
•Methods include K-Folding, bagging, boosting, and stacking.
• Diverse set of decision boundaries.
01 02
MNIST Dataset CNN for Classification
A Convolutional Neural Network (CNN) with
The study uses the MNIST dataset, consisting of
two convolutional layers and a fully connected
70,000 grayscale images of handwritten digits, to
layer is employed for the image classification
assess the defense against adversarial attacks.
task.
Adversarial Attack Generation and Defense
The K-fold technique is applied to the dataset The K-fold technique is applied to the dataset to train
to train and validate the autoencoder-based and validate the autoencoder-based defense, aiming
defense, aiming for a robust and generalizable for a robust and generalizable model.
model.
Results
01 Impact of Perturbations
Visualization of original, adversarial, and reconstructed images shows the deceptive nature of
adversarial attacks and the autoencoder's ability to recover the digit's structure.
The study compares the CNN model's classification accuracy on clean, adversarial, and
reconstructed test sets, demonstrating the autoencoder's effectiveness in improving robustness.
[2] Shekhar, S., et al. (2023). A Comprehensive Survey on Adversarial Attacks and Defenses in Machine Learning.
arXiv preprint arXiv:2312.03520.
[3] Goodfellow, I. J., et al. (2015). Explaining and harnessing adversarial examples. arXiv preprint
arXiv:1412.6572.
[4] Guo, C., et al. (2018). Countering adversarial images using input transformations. arXiv preprint
arXiv:1711.00117.
[5] Vincent, P., et al. (2008). Extracting and composing robust features with denoising autoencoders. In Proceedings
of the 25th international conference on Machine learning (pp. 1096-1103).
[6] Xu, W., Evans, D., Qi, Y.: Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks. arXiv
preprint arXiv:1704.01155 (2017).
[7] Samangouei, P., Kabkab, M., Chellappa, R.: Defense-GAN: Protecting Classifiers Against
Adversarial Attacks Using Generative Models. arXiv preprint arXiv:1805.06605 (2018).
THANK YOU