Description:

  • Without proper guardrails, generative AI outputs can contain confidential and/or sensitive information included in the model’s training dataset, RAG data sources, or data residing in data sources that the AI system is connected to (e.g., through language model tools such as agents or plugins). Examples of such information include that which is covered under data protection laws and regulations (e.g., personally identifiable information, protected health information, cardholder data) and corporate secrets.

Impact:

  • Can lead to a confidentiality breach of sensitive/confidential data included in the model’s training dataset, RAG data sources, or data residing in data sources that the AI system is connected to (e.g., through language model tools such as agents or plugins).
  • Can also lead to a confidentiality breach of the system’s metaprompt, which can be an extremely valuable piece intellectual property. Discovering the metaprompt can inform adversaries about the internal workings of the AI application and overall AI system.

Applies to which types of AI models? Generative AI specifically

Which AI security requirements function against this threat? [?]
Discussed in which authoritative sources? [?]
Discussed in which commercial sources? [?]