Prompt injection refers to a technique where users input specific prompts or instructions to influence the responses generated by a language model like ChatGPT.
However, threat actors mainly use this technique to mod the ChatGPT instances for several malicious purposes. It has several negative impacts like:-
An independent security researcher, Utku Sen, recently developed and launched a new tool dubbed “promptmap” that will enable users to test the prompt injection attacks on ChatGPT instances.
Promptmap
On ChatGPT instances, the “promptmap” automatically tests the prompt injections by understanding the context and purpose of your rules configured on ChatGPT.
It uses this understanding to create custom attack prompts for the target, running them alongside your system prompts. While this tool checks for prompt injection success by analyzing the ChatGPT instance’s response.
Work Mechanism Structure (Source – GitHub)
Attack types
Here below, we have mentioned all the current attack types along with their details:-
Support the originator by clicking the read the rest link below.