Breaking Trust with Words: Prompt Injection Leading to Simulated /etc/passwd Disclosure - resecurity.com
Prompt injection is a critical vulnerability in Large Language Models (LLMs) that manipulates their natural language processing to override system instructions and safety filters using crafted malicious prompts. This exploit can lead to significant impacts such as data exfiltration, including the disclosure of sensitive files like `/etc/passwd`, and enables the AI system to perform unauthorized actions.
Source: Original Report ↗