Poisoning AI Models: New Frontiers in Data Manipulation Attacks
2023, Poisoning AI Models: New Frontiers in Data Manipulation Attacks
Abstract
Artificial intelligence (AI) and data science models play a crucial role in critical sectors such as cybersecurity, healthcare, and finance, driving key insights and decision-making processes. However, as AI adoption grows, so does its exposure to emerging threats, particularly model poisoning attacks. In these attacks, adversaries stealthily manipulate training data to corrupt model behavior, either causing it to produce malicious outputs or rendering it ineffective against specific threats. This paper examines the methods and motivations behind data poisoning attacks, focusing on how adversaries compromise data pipelines, manipulate model performance, and evade detection. We also examine potential countermeasures and discuss ongoing research challenges that must be overcome to protect AI systems from these evolving threats.
References (28)
- B. Biggio, B. Nelson, and P. Laskov, "Poisoning attacks against support vector machines," in *Proc. Int. Conf. Mach. Learn. (ICML)*, 2012.
- T. Gu, K. Liu, B. Dolan-Gavitt, and S. Garg, "BadNets: Identifying vulnerabilities in the machine learning model supply chain," in *Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR) Workshops*, 2019.
- L. Huang et al., "Adversarial machine learning," *IEEE Trans. Knowl. Data Eng.*, vol. 23, no. 6, pp. 868-883, 2011.
- E. Bagdasaryan et al., "Poisoning attacks against federated learning systems," in *Proc. 33rd Int. Conf. Neural Inf. Process. Syst. (NeurIPS)*, 2020.
- M. A. Ilyas, L. A. S. Z. Shou, and B. Recht, "Poisoning attacks in machine learning: A survey," *IEEE Trans. Knowl. Data Eng.*, vol. 32, no. 2, pp. 178-189, 2020.
- C. Li, H. Xie, and W. Lou, "Secure machine learning for data poisoning attacks: A survey," *Int. J. Comput. Sci. Inf. Secur.*, vol. 16, no. 4, pp. 1-14, 2018.
- S. O. Moosavi-Dezfooli, A. Fawzi, and P. Frossard, "Analysis of adversarial attacks and defenses in machine learning," *IEEE Trans. Neural Netw. Learn. Syst.*, vol. 30, no. 9, pp. 2674-2687, 2019.
- A. B. Rad and A. M. Kermani, "Detecting adversarial samples in machine learning using model behavior analysis," in *Proc. IEEE Int. Conf. Mach. Learn. Appl. (ICMLA)*, 2018, pp. 45-51.
- M. S. Abdelkader, N. P. P. P. Biswas, and T. E. Moore, "Blockchain-based data provenance and its applications in machine learning,", vol. 8, pp. 130436-130445, 2020.
- J. Pan, Q. Yang, and P. S. Yu, "A survey of transfer learning," *IEEE Trans. Knowl. Data Eng.*, vol. 22, no. 10, pp. 1345-1359, 2010.
- C. Xiao et al., "Generating adversarial examples with adversarial networks," in *Proc. 27th Int. Joint Conf. Artif. Intell. (IJCAI)*, 2018, pp. 3905-3911.
- E. Bagdasaryan, A. Veit, Y. Hua, D. Estrin, and V. Shmatikov, "How to backdoor federated learning," in *Proc. 23rd Int. Conf. Artif. Intell. Stat. (AISTATS)*, 2020, pp. 2938-2948.
- A. Shafahi et al., "Poison frogs! Targeted clean-label poisoning attacks on neural networks," in *Adv. Neural Inf. Process. Syst. (NeurIPS)*, 2018, pp. 6103-6113.
- B. Biggio et al., "Evasion attacks against machine learning at test time," in *Proc. Eur. Conf. Mach. Learn. Princ. Pract. Knowl. Discov. Databases (ECML PKDD)*, 2013.
- M. Jagielski et al., "Manipulating machine learning: Poisoning attacks and countermeasures for regression learning," in *Proc. IEEE Symp. Secur. Privacy (SP)*, 2018.
- B. Biggio, G. Fumera, and F. Roli, "Pattern recognition systems under attack: Design issues and research challenges," *Int. J. Pattern Recognit. Artif. Intell.*, vol. 28, no. 07, p. 1460002, 2014.
- N. Papernot, P. McDaniel, A. Sinha, and M. P. Wellman, "SoK: Security and privacy in machine learning," in *Proc. IEEE Eur. Symp. Secur. Privacy (EuroS&P)*, 2018, pp. 399-414.
- Y. Liu, X. Chen, C. Liu, and D. Song, "Trojaning attack on neural networks," in *Proc. 25th Netw. Distrib. Syst. Secur. Symp. (NDSS)*, 2018.
- A. Steinhardt, P. W. Koh, and P. Liang, "Certified defenses for data poisoning attacks," in *Adv. Neural Inf. Process. Syst. (NeurIPS)*, 2017.
- S. G. Finlayson et al., "Adversarial attacks on medical machine learning," *Science*, vol. 363, no. 6433, pp. 1287- 1289, 2019.
- X. Ma, Y. Liu, J. Bailey, J. Lu, and Y. Jiang, "Understanding adversarial attacks on deep learning-based medical image analysis systems," *Pattern Recognit.*, vol. 110, p. 107332, 2020.
- M. Paschali, S. Conjeti, F. Navarro, and N. Navab, "Generalizability vs. robustness: Investigating medical imaging networks using adversarial examples," in *Proc. Med. Image Comput. Comput.-Assist. Interv. (MICCAI)*, 2018, pp. 493-501.
- E. Battista, B. Biggio, and F. Roli, "Adversarial machine learning in credit scoring: Threats and countermeasures," *IEEE Trans. Neural Netw. Learn. Syst.*, vol. 33, no. 7, pp. 3015-3028, 2022.
- Z. Yang, Y. Wang, J. Zhang, and X. Li, "Data poisoning attacks on financial fraud detection: A survey and empirical study," in *Proc. Int. Conf. Inf. Secur. Cryptol. (ICISC)*, 2021, pp. 214-230.
- W. Xu, D. Evans, and Y. Qi, "Feature squeezing: Detecting adversarial examples in deep neural networks," in *Proc.
- Netw. Distrib. Syst. Secur. Symp. (NDSS)*, 2019, pp. 1-15.
- B. I. P. Rubinstein, B. Nelson, L. Huang, A. D. Joseph, and J. D. Tygar, "Stealthy poisoning attacks on PCA-based anomaly detectors," in *Proc. 15th ACM SIGKDD Int. Conf. Knowl. Discov. Data Min. (KDD)*, 2009, pp. 435-444.
- R. Sommer and V. Paxson, "Outside the closed world: On using machine learning for network intrusion detection," in *Proc. IEEE Symp. Secur. Privacy*, 2010, pp. 305-316.