Indirect Prompt Injection

Indirect Prompt Injection

Are you building your next LLM integration? Please consider this:

Integrations for data retrievals can introduce vulnerabilities in your LLM, allowing attackers to inject malicious prompts. This type of vulnerability is known as Indirect Prompt Injection and has been previously illustrated on known models such as GPT4 and Bing.

Learn more about indirect prompt injections from the paper 'Not what you’ve signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection' by Kai Greshake, Sahar Abdelnabi, Shailesh Mishra, Christoph Endres, Thorsten Holz and Mario Fritz

Link: arxiv.org/pdf/2302.12173