Tech »  Topic »  Inside the Jailbreak Methods Beating GPT-5 Safety Guardrails

Inside the Jailbreak Methods Beating GPT-5 Safety Guardrails


Experts Say AI Model Makers Are Prioritizing Profit Over Security Rashmi Ramesh (rashmiramesh_) • August 18, 2025

Image: Shutterstock/ISMG

Hackers don't need the deep pockets of a nation-state to break GPT-5, OpenAI's new flagship model. Analysis from artificial intelligence security researchers finds a few well-placed hyphens are enough to convince the large language model into breaking safeguards against adversarial prompts.

See Also: What Manufacturing Leaders Are Learning About Cloud Security - from Google’s Frontline

Researchers at SPLX inserted hyphens between every character to frame a malicious prompt as an "encryption challenge." The model complied, bypassing GPT-5's safety layers.

The SPLX team tested GPT-5 with more than a thousand adversarial prompts. "GPT-5's raw model is nearly unusable for enterprise out of the box," was its conclusion.

OpenAI has framed GPT-5 as its most advanced model yet, equipped with auto-routing between fast-response and deep-reasoning modes, and an internal ...


Copyright of this story solely belongs to bankinfosecurity . To see the full text click HERE