Prompt injection: Difference between revisions
From Rice Wiki
(Created page with "Category:Cybersecurity A '''prompt injection''' attack involves a user injecting a malicious instruction in an LLM-integrated application, in which user input was intended to act as only data.") |
No edit summary |
||
Line 4: | Line 4: | ||
instruction in an LLM-integrated application, in which user input was | instruction in an LLM-integrated application, in which user input was | ||
intended to act as only data. | intended to act as only data. | ||
= Defense strategies = | |||
* [[StruQ]] rejects all user instructions | |||
* [[Instruction hierarchy]] rejects user instructions that are misaligned with the system prompt |
Revision as of 20:25, 23 May 2024
A prompt injection attack involves a user injecting a malicious
instruction in an LLM-integrated application, in which user input was
intended to act as only data.
Defense strategies
- StruQ rejects all user instructions
- Instruction hierarchy rejects user instructions that are misaligned with the system prompt