Back to articles
Anthropic Just Admitted Their New Model Is Too Dangerous to Release
How-ToSecurity

Anthropic Just Admitted Their New Model Is Too Dangerous to Release

via Dev.toAamer Mihaysi

Claude Mythos finds vulnerabilities that have existed for 27 years. Anthropic is only giving it to security researchers. This is the first real AI safety inflection point. Anthropic didn't release a model today. They announced Project Glasswing instead. The model, Claude Mythos, exists. It works. But you can't use it. Only a handful of security researchers get access. Why? Because it found a 27-year-old bug in OpenBSD by sending "a couple pieces of data to any server and crash it." Because it chains four or five vulnerabilities together to build exploits that escape sandboxes. Because it achieved a 181/200 success rate on exploit development where the previous best model managed 2. This is the first time an AI company has said: we built something we can't give you. The Capability Gap Is Real Let me put some numbers on this. Claude Opus 4.6: near-0% success rate at autonomous exploit development. Claude Mythos: 181 working exploits out of 200 attempts on Firefox JavaScript engine vulner

Continue reading on Dev.to

Opens in a new tab

Read Full Article
0 views

Related Articles