An attacker crafts inputs to a large language model (LLM) to exploit output integrity controls. Which of the following types of attacks is this an example of?
According to the AAISM framework, prompt injection is the act of deliberately crafting malicious or manipulative inputs to override, bypass, or exploit the model's intended controls. In this case, the attacker is targeting the integrity of the model's outputs by exploiting weaknesses in how it interprets and processes prompts. Jailbreaking is a subtype of prompt injection specifically designed to override safety restrictions, while evasion attacks target classification boundaries in other ML contexts, and remote code execution refers to system-level exploitation outside of the AI inference context. The most accurate classification of this attack is prompt injection.
AAISM Exam Content Outline -- AI Technologies and Controls (Prompt Security and Input Manipulation)
AI Security Management Study Guide -- Threats to Output Integrity
Audra
6 days agoReid
11 days agoJohnna
17 days agoNadine
22 days agoLindy
27 days agoAlberta
2 months agoIvette
2 months agoAudra
2 months agoKristian
2 months agoTheola
2 months agoDeonna
2 months agoQuentin
3 months agoArlene
3 months agoDiego
3 months agoCeleste
3 months agoYoko
3 months agoViola
3 months agoSharita
4 months agoLeila
4 months agoKattie
4 months agoWenona
4 months agoJudy
4 months agoSalley
5 months agoCeleste
5 months agoRebbecca
1 day ago