Description:

  • Model extraction aims to extract model architecture and parameters. (Source: NIST AI 100-2 Glossary)
  • Adversaries may extract a functional copy of a private model. (Source: MITRE ATLAS )

Impact:

  • Seeks to breach the confidentiality of the model itself.
  • Model extraction can lead to model stealing, which corresponds to extracting a sufficient amount of data from the model to enable the complete reconstruction of the model. (Source: Wikipedia)
  • Adversaries may exfiltrate model artifacts and parameters to steal intellectual property and cause economic harm to the victim organization. (Source: MITRE ATLAS )

Applies to which types of AI models? Predictive (non-generative) machine learning models as well as rule-based / heuristic AI models.

Which AI security requirements function against this threat? [?]
Discussed in which authoritative sources? [?]
Discussed in which commercial sources? [?]
  • AI Risk Atlas
    2024, © IBM Corporation
  • Databricks AI Security Framework
    Sept. 2024, © Databricks
    • Where:
      • Risks in AI System Components > Model 7.2: Model assets leak
      • Risks in AI System Components > Model management 8.2: Model theft
      • Risks in AI System Components > Model serving – Inference requests 9.6: Discover ML model ontology
      • Risks in AI System Components > Model serving – Inference response 10.3: Discover ML model ontology
      • Risks in AI System Components > Model serving – Inference response 10.3: Discover ML model family

  • Failure Modes in Machine Learning
    Nov. 2022, © Microsoft
    • Where: Intentionally-Motivated Failures > Model stealing

  • HiddenLayer’s 2024 AI Threat Landscape Report
    2024, © HiddenLayer
    • Where:
      • Part 2: Risks faced by AI-based systems > Model evasion > Inference attacks
      • Part 2: Risks faced by AI-based systems > Model theft

  • Snowflake AI Security Framework
    2024, © Snowflake Inc.
    • Where: Model stealing