Getting Started with AI Hacking Part 2: Prompt Injection - Black Hills Information Security, Inc.
Prompt injection is a critical vulnerability within Large Language Models (LLMs) that allows attackers to manipulate models into ignoring or overriding their original system instructions. This exploit enables LLMs to disclose sensitive information, bypass safety guidelines, or execute unintended actions by providing crafted input that redefines the model's behavior.
Source: Original Report ↗