Comprehensive and Detailed Explanation From Exact AWS AI documents:
This scenario describes prompt injection, which is a well-documented security and safety risk in generative AI systems.
Prompt injection occurs when a user intentionally crafts input prompts to override, manipulate, or bypass system instructions, guardrails, or safety policies defined by the AI application developer. The user’s instruction explicitly attempts to override prior system instructions and force the model into unrestricted behavior.
AWS Responsible AI and Generative AI security guidance describe prompt injection as:
An attempt to alter model behavior through malicious or manipulative user input
A risk that can lead to policy violations, unsafe outputs, or data misuse
A key concern when deploying large language models (LLMs) in production systems
Why the other options are incorrect:
Data bias (B) refers to skewed or unrepresentative training data, not user manipulation at inference time.
Hallucination (C) refers to the model generating incorrect or fabricated information.
Data exposure (D) involves leaking sensitive or private data, not instruction hijacking.
AWS AI document references (for exact extracts):
AWS Responsible AI Overview — section on Generative AI risks
Amazon Bedrock Security Best Practices — section on prompt injection and input validation
AWS Generative AI Governance Guidance — discussion of instruction hierarchy and guardrails