Abstract
Machine learning techniques, including Deep Reinforcement Learning (DRL), enhance intrusion detection systems by adapting to new threats. However, DRL’s reliance on vulnerable deep neural networks leads to susceptibility to adversarial examples-perturbations designed to evade detection. While adversarial examples are well-studied in deep learning, their impact on DRL-based intrusion detection remains underexplored, particularly in critical domains. This article conducts a thorough analysis of DRL-based intrusion detection’s vulnerability to adversarial examples. It systematically evaluates key hyperparameters such as DRL algorithms, neural network depth, and width, impacting agents’ robustness. The study extends to black-box attacks, demonstrating adversarial transferability across DRL algorithms. Findings emphasize neural network architecture’s critical role in DRL agent robustness, addressing underfitting and overfitting challenges. Practical implications include insights for optimizing DRL-based intrusion detection agents to enhance performance and resilience. Experiments encompass multiple DRL algorithms tested on three datasets: NSL-KDD, UNSW-NB15, and CICIoV2024, against gradient-based adversarial attacks, with publicly available implementation code.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Abou El Houda, Z., Moudoud, H., Brik, B.: Federated deep reinforcement learning for efficient jamming attack mitigation in o-ran. In: IEEE Transactions on Vehicular Technology (2024)
Ajay, J.: Why Cybersecurity Should Be Top Of Mind In 2023 (2023). https://www.forbes.com/sites/forbestechcouncil/2023/01/06/why-cybersecurity-should-be-top-of-mind-in-2023/?sh=1654d131235c
Annachhatre, C., Austin, T.H., Stamp, M.: Hidden Markov models for malware classification. J. Comput. Virol. Hack. Tech. 11, 59–73 (2015)
Bajaj, K., Arora, A.: Improving the intrusion detection using discriminative machine learning approach and improve the time complexity by data mining feature selection methods. Int. J. Comput. Appl. 76(1), 5–11 (2013)
Behzadan, V., Munir, A.: Vulnerability of Deep Reinforcement Learning to Policy Induction Attacks. In: Machine Learning and Data Mining in Pattern Recognition (2017)
Buczak, A.L., Guven, E.: A survey of data mining and machine learning methods for cyber security intrusion detection. IEEE Commun. Surv. Tutor. 18(2), 1153–1176 (2016)
Caminero, G., Lopez-Martin, M., Carro, B.: Adversarial environment reinforcement learning algorithm for intrusion detection. Comput. Netw. 159, 96–109 (2019)
Cevallos M., J.F., Rizzardi, A., Sicari, S., Coen Porisini, A.: Deep Reinforcement Learning for intrusion detection in Internet of Things: Best practices, lessons learnt, and open challenges. Computer Networks (2023)
Chebrolu, S., Abraham, A., Thomas, J.P.: Feature deduction and ensemble design of intrusion detection systems. Computers & Security (2005)
Chen, T., Liu, J., Xiang, Y., Niu, W., Tong, E., Han, Z.: Adversarial attack and defense in reinforcement learning-from ai security view. Cybersecurity (2019)
Dabney, W., Rowland, M., Bellemare, M., Munos, R.: Distributional reinforcement learning with quantile regression. In: Proceedings of the AAAI Conference on Artificial Intelligence (2018)
Debar, H., Becker, M., Siboni, D.: A neural network component for an intrusion detection system. In: Proceedings 1992 IEEE Computer Society Symposium on Research in Security and Privacy (1992)
Farebrother, J., Machado, M.C., Bowling, M.: Generalization and Regularization in DQN. In: NeurIPS 2018: Deep Reinforcement Learning Workshop (2018)
Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. In: 3rd International Conference on Learning Representations (2015)
Hasselt, H.v., Guez, A., Silver, D.: Deep reinforcement learning with double Q-Learning. In: AAAI Conference on Artificial Intelligence (2016)
He, K., Kim, D.D., Asghar, M.R.: Adversarial machine learning for network intrusion detection systems: a comprehensive survey. IEEE Commun. Surv. Tutor. 25(1), 538–566 (2023)
Hessel, M., Modayil, J., van Hasselt, H., Schaul, T., Ostrovski, G., Dabney, W., Horgan, D., Piot, B., Azar, M., Silver, D.: Rainbow: combining improvements in deep reinforcement learning. In: AAAI Conference on Artificial Intelligence (2018)
Hickling, T., Aouf, N., Spencer, P.: Robust adversarial attacks detection based on explainable deep reinforcement learning for uav guidance and planning. In: IEEE Transactions on Intelligent Vehicles (2023)
Huang, S., Papernot, N., Goodfellow, I., Duan, Y., Abbeel, P.: Adversarial Attacks on Neural Network Policies (2017). http://arxiv.org/abs/1702.02284
Ilahi, I., Usama, M., Qadir, J., Janjua, M.U., Al-Fuqaha, A., Hoang, D.T., Niyato, D.: Challenges and countermeasures for adversarial attacks on deep reinforcement learning. IEEE Trans. Artif. Intell. 3(2), 90–109 (2021)
Khraisat, A., Gondal, I., Vamplew, P., Kamruzzaman, J.: Survey of intrusion detection systems: techniques, datasets and challenges. Cybersecurity 2(1), 1–22 (2019)
Kiran, B.R., Sobh, I., Talpaert, V., Mannion, P., Sallab, A.A.A., Yogamani, S., Pérez, P.: Deep Reinforcement Learning for Autonomous Driving: A Survey (2021). http://arxiv.org/abs/2002.00444
Kos, J., Song, D.: Delving into adversarial attacks on deep policies. In: 5th International Conference on Learning Representations (2019)
Kreibich, C., Crowcroft, J.: Honeycomb: creating intrusion detection signatures using honeypots. ACM SIGCOMM Comput. Commun. Rev. 34(1), 51–56 (2004)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems. Curran Associates, Inc. (2012)
Kurakin, A., Goodfellow, I.J., Bengio, S.: Adversarial examples in the physical world. In: 5th International Conference on Learning Representations (2017)
Levine, S., Finn, C., Darrell, T., Abbeel, P.: End-to-End Training of Deep Visuomotor Policies (2016). http://arxiv.org/abs/1504.00702
Liao, H.J., Richard Lin, C.H., Lin, Y.C., Tung, K.Y.: Intrusion detection system: a comprehensive review. J. Netw. Comput. Appl. 36(1), 16–24 (2013)
Lillicrap, T.P., Hunt, J.J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., Wierstra, D.: Continuous control with deep reinforcement learning (2016). http://arxiv.org/abs/1509.02971
Lin, Y.C., Hong, Z.W., Liao, Y.H., Shih, M.L., Liu, M.Y., Sun, M.: Tactics of Adversarial Attack on Deep Reinforcement Learning Agents. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence (2017)
Lopez-Martin, M., Carro, B., Sanchez-Esguevillas, A.: Application of deep reinforcement learning to intrusion detection for supervised problems. Expert Syst. Appl. 141, 112963 (2020)
Meng, M.H., Bai, G., Teo, S.G., Hou, Z., Xiao, Y., Lin, Y., Dong, J.S.: Adversarial Robustness of deep neural networks: a survey from a formal verification perspective. IEEE Transactions on Dependable and Secure Computing (2022)
Merzouk, M.A., Cuppens, F., Boulahia-Cuppens, N., Yaich, R.: A Deeper analysis of adversarial examples in intrusion detection. In: 15th International Conference on Risks and Security of Internet and Systems (2021)
Merzouk, M.A., Cuppens, F., Boulahia-Cuppens, N., Yaich, R.: Investigating the practicality of adversarial evasion attacks on network intrusion detection. Anna. Telecommun. 77(11), 763–775 (2022)
Mnih, V., Badia, A.P., Mirza, M., Graves, A., Lillicrap, T., Harley, T., Silver, D., Kavukcuoglu, K.: Asynchronous Methods for Deep Reinforcement Learning. In: Proceedings of The 33rd International Conference on Machine Learning (2016)
Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller, M.: Playing atari with deep reinforcement learning. http://arxiv.org/abs/1312.5602 (2013)
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M., Fidjeland, A.K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., Hassabis, D.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)
Mohamed, S., Ejbali, R.: Deep SARSA-based reinforcement learning approach for anomaly network intrusion detection system. Int. J. Inf. Security 22(1), 235–247 (2023)
Moudoud, H., Cherkaoui, S.: Empowering security and trust in 5g and beyond: A deep reinforcement learning approach. IEEE Open Journal of the Communications Society (2023)
Moustafa, N., Slay, J.: UNSW-NB15: a comprehensive data set for network intrusion detection systems (UNSW-NB15 network data set). In: Military Communications and Information Systems Conference (MilCIS) (2015)
Neto, E.C.P., Taslimasa, H., Dadkhah, S., Iqbal, S., Xiong, P., Rahman, T., Ghorbani, A.A.: CICIoV2024: advancing realistic IDS approaches against DoS and spoofing attack in IoV CAN bus. Internet Things 26, 101209 (2024)
Nicolae, M.I., Sinn, M., Tran, M.N., Buesser, B., Rawat, A., Wistuba, M., Zantedeschi, V., Baracaldo, N., Chen, B., Ludwig, H., Molloy, I.M., Edwards, B.: Adversarial Robustness Toolbox v1.0.0. http://arxiv.org/abs/1807.01069 (2019)
Papernot, N., McDaniel, P., Goodfellow, I.: Transferability in machine learning: from phenomena to black-box attacks using adversarial samples (2016). http://arxiv.org/abs/1605.07277
Papernot, N., McDaniel, P., Goodfellow, I., Jha, S., Celik, Z.B., Swami, A.: Practical black-box attacks against machine learning. In: Proceedings of the ACM on Asia Conference on Computer and Communications Security (2017)
Papernot, N., McDaniel, P., Jha, S., Fredrikson, M., Celik, Z.B., Swami, A.: The limitations of deep learning in adversarial settings. In: IEEE European Symposium on Security and Privacy (2016)
Pasikhani, A.M., Clark, J.A., Gope, P.: Adversarial RL-based IDS for evolving data environment in 6LoWPAN. IEEE Trans. Inf. Forens. Security 17, 3831–3846 (2022)
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: Pytorch: An imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems (2019)
Pierazzi, F., Pendlebury, F., Cortellazzi, J., Cavallaro, L.: Intriguing Properties of Adversarial ML Attacks in the Problem Space. In: IEEE Symposium on Security and Privacy (SP) (2020)
Priya, S., Kumar, K.P.M.: Binary bat algorithm based feature selection with deep reinforcement learning technique for intrusion detection system. Soft Comput. 27(15), 10777–10788 (2023)
Raffin, A., Hill, A., Gleave, A., Kanervisto, A., Ernestus, M., Dormann, N.: Stable-baselines3: reliable reinforcement learning implementations. J. Mach. Learn. Res. 22(268), 1–8 (2021)
Rice, L., Wong, E., Kolter, Z.: Overfitting in adversarially robust deep learning. In: Proceedings of the 37th International Conference on Machine Learning (2020)
Ring, M., Wunderlich, S., Scheuring, D., Landes, D., Hotho, A.: A survey of network-based intrusion detection data sets. Comput. Security 86, 147–167 (2019)
Schulman, J., Levine, S., Abbeel, P., Jordan, M., Moritz, P.: Trust region policy optimization. In: International conference on machine learning (2015)
Schulman, J., Moritz, P., Levine, S., Jordan, M.I., Abbeel, P.: High-dimensional continuous control using generalized advantage estimation (2016). http://arxiv.org/abs/1506.02438
Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms. http://arxiv.org/abs/1707.06347 (2017)
Servin, A., Kudenko, D.: Multi-agent Reinforcement Learning for Intrusion Detection. In: Adaptive Agents and Multi-Agent Systems III. Adaptation and Multi-Agent Learning (2008)
Sewak, M., Sahay, S.K., Rathore, H.: Deep reinforcement learning in the advanced cybersecurity threat detection and protection. Inf. Syst. Front. 25(2), 589–611 (2023)
Sheatsley, R., Papernot, N., Weisman, M.J., Verma, G., McDaniel, P.: Adversarial examples for network intrusion detection systems. J. Comput. Security 30, 727 (2022)
Sun, J., Zhang, T., Xie, X., Ma, L., Zheng, Y., Chen, K., Liu, Y.: Stealthy and efficient adversarial attacks against deep reinforcement learning. In: AAI Conference on Artificial Intelligence (2020)
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems. Curran Associates, Inc. (2014)
Sutton, R.S.: Learning to predict by the methods of temporal differences. Mach. Learn. 3, 9–44 (1988)
Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., Fergus, R.: Intriguing properties of neural networks. In: 2nd International Conference on Learning Representations, ICLR (2014)
Tavallaee, M., Bagheri, E., Lu, W., Ghorbani, A.A.: A detailed analysis of the kdd cup 99 data set. In: IEEE symposium on computational intelligence for security and defense applications (2009)
Umer, M.A., Junejo, K.N., Jilani, M.T., Mathur, A.P.: Machine learning for intrusion detection in industrial control systems: applications, challenges, and recommendations. Int. J. Critic. Infrastruct. Protect. 38, 100516 (2022)
Wang, Z.: Deep learning-based intrusion detection with adversaries. IEEE Access 6, 38367–38384 (2018)
Wang, Z., Schaul, T., Hessel, M., Van Hasselt, H., Lanctot, M., De Freitas, N.: Dueling network architectures for deep reinforcement learning. In: 33rd International Conference on International Conference on Machine Learning - Volume 48 (2016)
Xiong, Z., Eappen, J., Zhu, H., Jagannathan, S.: Defending observation attacks in deep reinforcement learning via detection and denoising. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 235–250 (2022)
Xu, X., Xie, T.: A Reinforcement Learning Approach for Host-Based Intrusion Detection Using Sequences of System Calls. In: Advances in Intelligent Computing (2005)
Yilmaz, I., Masum, R., Siraj, A.: Addressing imbalanced data problem with generative adversarial network for intrusion detection. In: IEEE 21st International Conference on Information Reuse and Integration for Data Science (2020)
Yuan, X., He, P., Zhu, Q., Li, X.: Adversarial examples: attacks and defenses for deep learning. IEEE Trans. Neural Netw. Learn. Syst. 30(9), 2805–2824 (2019)
Acknowledgements
This work was supported by Mitacs through the Mitacs Accelerate International program and the CRITiCAL chair. It was enabled in part by support provided by Calcul Québec, Compute Ontario, the BC DRI Group, and the Digital Research Alliance of Canada.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors have no conflict of interest to declare.
Ethical approval
The authors confirm their compliance with ethical standards.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
IoV case study (CICIoV2024)
IoV case study (CICIoV2024)
In this section, we apply our method to the CICIoV2024 dataset. In contrast with the two previous datasets, NSL-KDD and UNSW-NB15, which contain data from TCP/IP networks, CICIoV2024 consists of IoV data captured from real car components. Here we demonstrate the applicability of our method to a more recent dataset and a different type of data. More precisely, we focus on the impact of the DRL algorithm used by the detection agent on the robustness. We use all 5 algorithms described in Sect. 3.1 with 4 hidden layer sizes (1, 2, 4, 8) and 4 hidden unit sizes (32, 64, 128, 256), with a total of 80 combinations.
Figure 6 shows a line plot of the FNR for each network width and depth combination. The lines correspond to different DRL algorithms (FGSM and BIM are represented with continuous and dashed lines, respectively). On the other hand, Fig. 5 shows the FNR mean and standard deviation averaged together over all neural network architectures, it is the summarized version of Fig. 6.
First, we observe the performance of DRL agents before adversarial perturbations (\(\epsilon =0\)). Figure 6 shows an FNR equal to 0 for almost all DRL algorithms on all architectures. Our results also show an F1 score equal to 1, demonstrating a perfect classification. These results are due to the relatively simple structure of the CICIoV2024 dataset compared to network datasets, the former consists of 8 values each encoded in one byte and an ID. IoV components are designed to be lightweight and energy-efficient, limiting the size of data communication to the minimum.
Similarly to Fig. 3, Figure 5 shows some variance because the values are averaged over 16 different architectures. We can nonetheless observe the behavioral patterns of DRL algorithms on CICIoV2024. First, we notice that the properties of the datasets disadvantage A2C which ends up with the highest average FNR, reaching an average around 0.8 at \(\epsilon =0.1\). Similarly to previous datasets, DQN, QRDQN, and PPO show comparable performance in terms of robustness. Their FNR evolves similarly and reaches 0.5 to 0.6 when \(\epsilon =0.1\). Finally, TRPO maintains the best robustness to adversarial examples, especially against FGSM, with a low FNR below \(\epsilon =0.6\) and an average FNR of 0.37.
Furthermore, the detailed Fig. 6 shows the specific behavior of the algorithms with different neural network architectures. It shows the relative stability of A2C across architectures, with an FNR increasing starting from around \(\epsilon =0.04\). PPO and TRPO show a similar behavior overall, in addition to a decrease in FNR with larger neural networks (especially 8 hidden layers). Moreover, DQN and QRDQN show a higher sensitivity to adversarial examples on smaller neural networks (up to 4 hidden layers and 64 units), with high FNR values starting from \(\epsilon =0.02\). The two algorithms are also the only ones that reach an FNR equal to 1 on some architectures. Finally, Regardless of the DRL algorithm used, agents with the largest neural network (8 hidden layers of 256 units) achieved the best robustness, with an FNR below 0.4 for all algorithms (except QRDQN against BIM).
In conclusion, the performance of DRL-based intrusion detection is not limited to the network domain, it extends to other critical tasks, and so do their vulnerabilities. In this section, we have shown the efficiency of DRL detection agents in detecting attacks against IoV devices using the CICIoV2024 dataset. Across different architectures, our agents present a perfect detection across almost all DRL algorithms and neural network architectures. These performances are explained by the relative simplicity and lightheartedness of IoV data. However, DRL agents also present similar vulnerabilities to adversarial attacks in IoV. We have shown how gradient-based adversarial examples with a perturbation amplitude \(\epsilon \le 0.1\) could considerably increase the FNR (attacks labeled as benign). This case study on IoV data demonstrates the seriousness of the adversarial examples threat in critical infrastructures and the need for further investigation on the robustness of DRL agents in IoV environments.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Merzouk, M.A., Neal, C., Delas, J. et al. Adversarial robustness of deep reinforcement learning-based intrusion detection. Int. J. Inf. Secur. 23, 3625–3651 (2024). https://doi.org/10.1007/s10207-024-00903-2
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10207-024-00903-2