Site icon VMVirtualMachine.com

Researchers Find Easy Way to Jailbreak Every Major AI, From ChatGPT to Claude

Researchers Find Easy Way to Jailbreak Every Major AI, From ChatGPT to Claude

Security researchers have discovered a highly effective new jailbreak that can dupe nearly every major large language model into producing harmful output, from explaining how to build nuclear weapons to encouraging self-harm.

As detailed in a writeup by the team at AI security firm HiddenLayer, the exploit is a prompt injection technique that can bypass the “safety guardrails across all major frontier AI models,” including Google’s Gemini 2.5, Anthropic’s Claude 3.7, and OpenAI’s…

Article Source
https://futurism.com/easy-jailbreak-every-major-ai-chatgpt

Exit mobile version