Illinois information sciences researchers develop AI safety testing methods - Illinois News Bureau
Researchers developed novel jailbreak methods, including "InfoFlood" and "JAMBench," to expose critical vulnerabilities in Large Language Model (LLM) moderation guardrails. These techniques successfully bypassed safety protocols, enabling LLMs to generate harmful content by exploiting input complexity and output filtering weaknesses.
Source: Original Report ↗