Inside the Jailbreak Methods Beating GPT-5 Safety Guardrails
bankinfosecurityExperts Say AI Model Makers Are Prioritizing Profit Over Security Rashmi Ramesh (rashmiramesh_) • August 18, 2025

Hackers don't need the deep pockets of a nation-state to break GPT-5, OpenAI's new flagship model. Analysis from artificial intelligence security researchers finds a few well-placed hyphens are enough to convince the large language model into breaking safeguards against adversarial prompts.
See Also: What Manufacturing Leaders Are Learning About Cloud Security - from Google’s Frontline
Researchers at SPLX inserted hyphens between every character to frame a malicious prompt as an "encryption challenge." The model complied, bypassing GPT-5's safety layers.
The SPLX team tested GPT-5 with more than a thousand adversarial prompts. "GPT-5's raw model is nearly unusable for enterprise out of the box," was its conclusion.
OpenAI has framed GPT-5 as its most advanced model yet, equipped with auto-routing between fast-response and deep-reasoning modes, and an internal ...
Copyright of this story solely belongs to bankinfosecurity . To see the full text click HERE