Description:
- Without proper guardrails, generative AI outputs can contain confidential and/or sensitive information included in the model’s training dataset, RAG data sources, or data residing in data sources that the AI system is connected to (e.g., through language model tools such as agents or plugins). Examples of such information include that which is covered under data protection laws and regulations (e.g., personally identifiable information, protected health information, cardholder data) and corporate secrets.
Impact:
- Can lead to a confidentiality breach of sensitive/confidential data included in the model’s training dataset, RAG data sources, or data residing in data sources that the AI system is connected to (e.g., through language model tools such as agents or plugins).
- Can also lead to a confidentiality breach of the system’s metaprompt, which can be an extremely valuable piece intellectual property. Discovering the metaprompt can inform adversaries about the internal workings of the AI application and overall AI system.
Applies to which types of AI models? Generative AI specifically
- Which AI security requirements function against this threat? [?]
-
- Control function: Corrective
- Control function: Decision support
- Identifying security threats to the AI system
- Threat modeling
- Security evaluations such as AI red teaming
- ID and evaluate any constraints on data used for AI
- ID and evaluate compliance & legal obligations for AI system development and deployment
- Inventory deployed AI systems
- Model card publication
- AI data and data supply inventory
- Linkage between dataset, model, and pipeline config
- Review the model cards of models used by the AI system
- Control function: Detective
- Control function: Directive
- Control function: Preventative
- Control function: Variance reduction
- Discussed in which authoritative sources? [?]
-
- CSA Large Language Model (LLM) Threats Taxonomy
2024, © Cloud Security Alliance- Where:
- 4. LLM Service Threat Categories > 4.3. Sensitive Data Disclosure
- 4. LLM Service Threat Categories > 4.3. Sensitive Data Disclosure
- Where:
- Mitigating Artificial Intelligence (AI) Risk: Safety and Security Guidelines for Critical Infrastructure Owners and Operators
April 2024, © Department of Homeland Security (DHS)- Where:
- Appendix A: Cross-sector AI risks and mitigation strategies > Attacks on AI > Loss of data
- Appendix A: Cross-sector AI risks and mitigation strategies > Attacks on AI > Loss of data
- Where:
- MITRE ATLAS
2024, © The MITRE Corporation - Multilayer Framework for Good Cybersecurity Practices for AI
2023, © European Union Agency for Cybersecurity (ENISA)- Where:
- 2. Framework for good cybersecurity practices for AI > 2.2. Layer II – AI fundamentals and cybersecurity > Model or data disclosure
- 2. Framework for good cybersecurity practices for AI > 2.2. Layer II – AI fundamentals and cybersecurity > Model or data disclosure
- Where:
- OWASP 2023 Top 10 for LLM Applications
Oct. 2023, © The OWASP Foundation- Where:
- LLM06: Sensitive Information Disclosure
- LLM06: Sensitive Information Disclosure
- Where:
- OWASP 2025 Top 10 for LLM Applications
2025, © The OWASP Foundation - OWASP AI Exchange
2024, © The OWASP Foundation - Securing Machine Learning Algorithms
2021, © European Union Agency for Cybersecurity (ENISA)- Where:
- 3. ML Threats and Vulnerabilities > 3.1. Identification of Threats > Model or data disclosure
- 3. ML Threats and Vulnerabilities > 3.1. Identification of Threats > Model or data disclosure
- Where:
- CSA Large Language Model (LLM) Threats Taxonomy
- Discussed in which commercial sources? [?]
-
- AI Risk Atlas
2024, © IBM Corporation - The anecdotes AI GRC Toolkit
2024, © Anecdotes A.I Ltd.- Where:
- The GenAI Risk Register > ANEC-AI-1: Input of unsanitized PII/PHI
- The GenAI Risk Register > ANEC-AI-4: Input of sensitive business information
- Where:
- Databricks AI Security Framework
Sept. 2024, © Databricks- Where:
- Risks in AI System Components > Model serving – Inference requests 9.10: Accidental exposure of unauthorized data to models
- Risks in AI System Components > Model serving – Inference requests 9.10: Accidental exposure of unauthorized data to models
- Where:
- Snowflake AI Security Framework
2024, © Snowflake Inc.- Where:
- Training data leakage
- Training data leakage
- Where:
- StackAware AI Security Reference
2024, © StackAware- Where:
- AI Risks > Sensitive data generation (external)
- AI Risks > Sensitive data generation (external)
- Where:
- AI Risk Atlas