Description: Model poisoning attacks attempt to directly modify the trained AI model to inject malicious functionality into the model. Once trained, a model is often just a file residing on a server. Attackers can alter the model file or replace it entirely with a poisoned model file. In this respect, even if a model has been correctly trained with a dataset that has been thoroughly vetted, this model can still be replaced with a poisoned model at various points in the AI SDLC or in the runtime environment.
Impact: Affects the integrity of model outputs, decisions, or behaviors.
Applies to which types of AI models? Any
- Which AI security requirements function against this threat? [?]
-
- Control function: Corrective
- Control function: Decision support
- Identifying security threats to the AI system
- Threat modeling
- Security evaluations such as AI red teaming
- ID and evaluate any constraints on data used for AI
- ID and evaluate compliance & legal obligations for AI system development and deployment
- Inventory deployed AI systems
- Model card publication
- Linkage between dataset, model, and pipeline config
- Review the model cards of models used by the AI system
- Control function: Detective
- Control function: Directive
- Control function: Preventative
- Control function: Resistive
- Control function: Variance reduction
- Discussed in which authoritative sources? [?]
-
- Attacking Artificial Intelligence: AI’s Security Vulnerability and What Policymakers Can Do About It
August 2019, Belfer Center for Science and International Affairs, Harvard Kennedy School- Where:
- Part I. Technical Problem > Poisoning Attacks > Model Poisoning
- Part I. Technical Problem > Poisoning Attacks > Model Poisoning
- Where:
- Cybersecurity of AI and Standardization
March 2023, © European Union Agency for Cybersecurity (ENISA)- Where:
- 4. Analysis of coverage > 4.1. Standardization in support of cybersecurity of AI – Narrow sense
- 4. Analysis of coverage > 4.1. Standardization in support of cybersecurity of AI – Narrow sense
- Where:
- MITRE ATLAS
2024, © The MITRE Corporation - Multilayer Framework for Good Cybersecurity Practices for AI
2023, © European Union Agency for Cybersecurity (ENISA)- Where:
- 2. Framework for good cybersecurity practices for AI > 2.2. Layer II – AI fundamentals and cybersecurity > Poisoning
- 2. Framework for good cybersecurity practices for AI > 2.2. Layer II – AI fundamentals and cybersecurity > Poisoning
- Where:
- NIST AI 100-2 E2023: Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations
Jan. 2024, National Institute of Standards and Technology (NIST)- Where:
- 2. Predictive AI Taxonomy > 2.3. Poisoning Attacks and Mitigations > 2.3.4. Model Poisoning
- 2. Predictive AI Taxonomy > 2.3. Poisoning Attacks and Mitigations > 2.3.4. Model Poisoning
- Where:
- OWASP Machine Learning Security Top 10
2023, © The OWASP Foundation - OWASP AI Exchange
2024, © The OWASP Foundation - OWASP 2025 Top 10 for LLM Applications
2025, © The OWASP Foundation - Securing Artificial Intelligence (SAI); AI Threat Ontology
2022, © European Telecommunications Standards Institute (ETSI)- Where:
- 6. Threat landscape > 6.4. Threat modeling > 6.4.2.3 > Implementation
- 6. Threat landscape > 6.4. Threat modeling > 6.4.2.3 > Implementation
- Where:
- Securing Machine Learning Algorithms
2021, © European Union Agency for Cybersecurity (ENISA)- Where:
- 3. ML Threats and Vulnerabilities > 3.1. Identification of Threats > Poisoning
- 3. ML Threats and Vulnerabilities > 3.1. Identification of Threats > Poisoning
- Where:
- Attacking Artificial Intelligence: AI’s Security Vulnerability and What Policymakers Can Do About It
- Discussed in which commercial sources? [?]
-
- Databricks AI Security Framework
Sept. 2024, © Databricks- Where:
- Risks in AI System Components > Algorithms 5.3: Hypermeters stealing
- Risks in AI System Components > Algorithms 7.1: Backdoor machine learning / Trojaned model
- Where:
- HiddenLayer’s 2024 AI Threat Landscape Report
2024, © HiddenLayer- Where:
- Part 2: Risks faced by AI-based systems > Malicious models
- Part 2: Risks faced by AI-based systems > Model backdoors
- Where:
- Snowflake AI Security Framework
2024, © Snowflake Inc.- Where:
- Backdooring models (insider attacks)
- Model poisoning
- Where:
- StackAware AI Security Reference
2024, © StackAware- Where:
- AI Risks > Corrupted model seeding (external)
- AI Risks > Corrupted model seeding (external)
- Where:
- Databricks AI Security Framework