Opus 4.6 uncovers 500 zero-day flaws in open-source code

https://www.axios.com/2026/02/05/anthropic-claude-opus-46-software-hunting

Exclusive: Anthropic's new model is a pro at finding security flaws

The AI company sees the model's advancements as a major win for cyber defenders in the race against adversarial AI.

Axios

The system card unfortunately only refers to this [0] blog post and doesn't go into any more detail. In the blog post Anthropic researchers claim: "So far, we've found and validated more than 500 high-severity vulnerabilities".

The three examples given include two Buffer Overflows which could very well be cherrypicked. It's hard to evaluate if these vulns are actually "hard to find". I'd be interested to see the full list of CVEs and CVSS ratings to actually get an idea how good these findings are.

Given the bogus claims [1] around GenAI and security, we should be very skeptical around these news.

[0] https://red.anthropic.com/2026/zero-days/

[1] https://doublepulsar.com/cyberslop-meet-the-new-threat-actor...

0-Days \ red.anthropic.com

I know some of the people involved here, and the general chatter around LLM-guided vulnerability discovery, and I am not at all skeptical about this.
[flagged]
It does if the person making the statement has a track record, proven expertise on the topic - and in this case… it actually may mean something to other people
Contra Ptacek's Terrible Article On AI — Ludicity

Here's a fun exercise: go email the author of that blog (he's very nice) and ask how much of it he still stands by.