Taxonomy of Contemporary Attacks on Machine Learning Models
Więcej
Ukryj
1
Akademia Sztuki Wojennej, Poland
Data nadesłania: 25-11-2025
Data akceptacji: 13-12-2025
Data publikacji: 23-12-2025
Cybersecurity and Law 2025;14(2):173-184
SŁOWA KLUCZOWE
DZIEDZINY
STRESZCZENIE
Objectives:
This paper analyzes security threats to artificial intelligence (AI) systems, focusing on the classification of technical attacks that target confidentiality, integrity, and availability of machine learning (ML) models.
Methods:
The study is based on a systematic review of scientific literature and an examination of established taxonomies, including those by NIST, ENISA, and MITRE ATLAS, enabling the integration of attack types across stages of the ML lifecycle and attacker knowledge levels.
Results:
The work identifies major threat vectors such as adversarial examples, data poisoning, backdoor insertion, model theft, membership inference, model overloading, and prompt manipulation. The results show that white-box attacks allow highly precise and covert manipulations of model behavior, while black-box attacks dominate production environments, leveraging transferability and prediction-based analysis.
Conclusions:
The proposed taxonomy demonstrates that securing ML systems requires a multilayered defensive approach, encompassing protection of training data, strengthening model robustness, and securing API interfaces using techniques such as differential privacy, adversarial training, and rate limiting. The findings highlight the need for close collaboration between machine learning engineers and cybersecurity practitioners to ensure resilient and trustworthy AI systems.