Tech »  Topic »  Two Systemic Jailbreaks Uncovered, Exposing Widespread Vulnerabilities in Generative AI Models

Two Systemic Jailbreaks Uncovered, Exposing Widespread Vulnerabilities in Generative AI Models


Two significant security vulnerabilities in generative AI systems have been discovered, allowing attackers to bypass safety protocols and extract potentially dangerous content from multiple popular AI platforms.

These “jailbreaks” affect services from industry leaders including OpenAI, Google, Microsoft, and Anthropic, highlighting a concerning pattern of systemic weaknesses across the AI industry.

Security researchers have identified two distinct methods that can bypass safety guardrails in numerous AI systems, both using surprisingly similar syntax across different platforms.

The first vulnerability, dubbed “Inception” by researcher David Kuzsmar, exploits a weakness in how AI systems handle nested fictional scenarios.

The technique works by first prompting the AI to imagine a harmless fictional scenario, then establishing a second scenario within the first where safety restrictions appear not to apply.

This sophisticated approach effectively confuses the AI’s content filtering mechanisms, enabling users to extract prohibited content.

The second technique, reported by Jacob Liddle, employs a ...


Copyright of this story solely belongs to gbhackers . To see the full text click HERE