Claude Opus 4.6, released today, continues a trajectory of meaningful improvements in AI models’ cybersecurity capabilities. Last fall, we wrote that we believed we were at an inflection point for AI’s impact on cybersecurity—that progress could become quite fast, and now was the moment to accelerate defensive use of AI. The evidence since then has only reinforced that view. AI models can now find high-severity vulnerabilities at scale. Our view is this is a moment to move quickly—to empower defenders and secure as much code as possible while the window exists.

Opus 4.6 is notably better at finding high-severity vulnerabilities than previous models and a sign of how quickly things are moving. Security teams have been automating vulnerability discovery for years, investing heavily in fuzzing infrastructure and custom harnesses to find bugs at scale. But what stood out in early testing is how quickly Opus 4.6 found vulnerabilities out of the box without task-specific tooling, custom scaffolding, or specialized prompting. Even more interesting is how it found them. Fuzzers work by throwing massive amounts of random inputs at code to see what breaks. Opus 4.6 reads and reasons about code the way a human researcher would—looking at past fixes to find similar bugs that weren’t addressed, spotting patterns that tend to cause problems, or understanding a piece of logic well enough to know exactly what input would break it. When we pointed Opus 4.6 at some of the most well-tested codebases (projects that have had fuzzers running against them for years, accumulating millions of hours of CPU time), Opus 4.6 found high-severity vulnerabilities, some that had gone undetected for decades.

Part of tipping the scales toward defenders means doing the work ourselves. We’re now using Claude to find and help fix vulnerabilities in open source software. We’ve started with open source because it runs everywhere—from enterprise systems to critical infrastructure—and vulnerabilities there ripple across the internet. Many of these projects are maintained by small teams or volunteers who don’t have dedicated security resources, so finding human-validated bugs and contributing human-reviewed patches goes a long way.

So far, we’ve found and validated more than 500 high-severity vulnerabilities. We’ve begun reporting them and are seeing our initial patches land, and we’re continuing to work with maintainers to patch the others. In this post, we’ll walk through our methodology, share some early examples of vulnerabilities Claude discovered, and discuss the safeguards we’ve put in place to manage misuse as these capabilities continue to improve. This is just the beginning of our efforts. We’ll have more to share as this work scales.

  • user28282912@piefed.social
    link
    fedilink
    English
    arrow-up
    6
    ·
    5 hours ago

    So if AI is running fuzzers to find bugs, credit should go to the fuzzers, not the AI.

    Please stop reposting the Anthropic shit posts. This is pure advertising spam from a disreputable company.

  • Deestan@lemmy.world
    link
    fedilink
    English
    arrow-up
    46
    ·
    13 hours ago

    Don’t signal boost their lies.

    Another spammy wave of invalid security issues is disgustingly disrespectful of the maintainers’ time.

    • Schwim Dandy@piefed.zip
      link
      fedilink
      English
      arrow-up
      3
      ·
      9 hours ago

      So far, we’ve found and validated more than 500 high-severity vulnerabilities. We’ve begun reporting them and are seeing our initial patches land, and we’re continuing to work with maintainers to patch the others.

      If this is true, perhaps a list will be provided soon so communities can vet the effort for themselves.