Anthropic’s new Claude Security tool scans your codebase for flaws – and helps you decide what to fix first


Claude Security

Elyse Betters Picaro / ZDNET

Follow ZDNET: Add us as a preferred source on Google.


ZDNET’s key takeaways

  • AI vulnerability scanning is moving into developer workflows.
  • Claude Security turns findings into prioritized fix guidance.
  • The big challenge is keeping these tools from attackers.

Anthropic has announced Claude Security, a new defensive cybersecurity product. Right now, it’s available in public beta to Enterprise-tier Claude users, with availability “coming soon” to Claude Team and Max-tier users.

Also: Apple, Google, and Microsoft join Anthropic’s Project Glasswing to defend world’s most critical software

Claude Security is another tool in Anthropic’s cyberdefense toolbox. It gives security teams a way to “scan codebases for vulnerabilities and generate targeted patches” using the Claude Opus 4.7 model.

Earlier in the month, Anthropic debuted Project Glasswing, an AI Manhattan Project aimed at finding vulnerabilities in the world’s infrastructure of open-source software.

Glasswing uses an Anthropic model called Mythos, a model deemed so dangerous that it’s not being released to the public. It’s being shared with Glasswing participants, including erstwhile competitors like Amazon Web Services, Anthropic, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorgan Chase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks.

Vulnerability scanning

At the core of both Project Glasswing and Claude Security is vulnerability scanning. Most cyberattacks begin with an enemy actor exploiting a vulnerability. So, if defenders can find and patch the vulnerabilities, the malicious perpetrator has a smaller attack surface.

Remember Star Wars? The entire plot of A New Hope revolves around Death Star plans that Princess Leia stores in R2-D2. Once the Rebels get those plans, they’re able to find a vulnerability. All Luke and the other pilots have to do is fire one torpedo down an exhaust port on the Death Star, and… boom!

That, boys and girls, is a vulnerability. The Death Star had one fatal flaw. Your codebase probably has more. Anthropic’s new Claude Security tool wants to find them before attackers get there first.

Back in the real world, everything runs on software, which is inherently vulnerable. Not only do vulnerabilities open doors for adversaries to exploit, but they also could cause damage simply by existing and causing bugs experienced by users of the software.

Also: I teamed up two AI tools to solve a major bug – but they couldn’t do it without me

I first used AI to do vulnerability scanning back in September with OpenAI’s Codex. At the time, it failed because it couldn’t handle a project-wide context. But when I teamed the AI pair programming tool with ChatGPT’s Deep Research, which was better with lots of data, the two found a number of critical vulnerabilities in my security software, which I immediately fixed.

Since then, both Codex and Claude Code have got better in terms of how much code they can process in one context, but neither is capable of handling an entire large codebase at once.

Mythos can, however. It can even handle the relationships between codebases on a macro scale. But it’s not available to the public, even via Enterprise-tier fees. Last month, OpenAI introduced Codex Security, which also offers a larger-scope context analysis. And now Claude Security can do similar larger-scale scans.

This new product is capable of scanning a full repository or a targeted directory. According to Anthropic, “Claude reasons about code the way a security researcher does, tracing data flows, reading source code, and working out how components interact across files and modules.”

There’s more to Claude Security, but first let’s talk about the big vulnerability introduced by vulnerability-scanning AIs.

Weapons of digital destruction

Vulnerability scanners help defenders defend. But they also help attackers find where to attack. That was the whole point with the Rebels’ attack on the Death Star. Once they knew of a vulnerability, they could exploit it.

For example, both Microsoft and OpenAI have reported that state-affiliated actors from China, Iran, Russia, and North Korea have used large language models to research various companies and cybersecurity tools, debug code, generate scripts, and create content likely for use in phishing and spear-phishing campaigns.

Also: AI is getting scary good at finding hidden software bugs – even in decades-old code

Anthropic is trying to prevent its models from being used in similar ways. As of the launch of Opus 4.7, the company includes new cyber safeguards that automatically detect and block requests suggestive of prohibited or high-risk cybersecurity uses.

For example, Opus 4.7 now blocks “Activities that are almost always used maliciously and have little to no legitimate defensive application such as mass data exfiltration or ransomware code development.”

On the other hand, what about activities that have legitimate defensive applications, such as vulnerability exploitation or offensive security tooling development? Opus 4.7 also blocks these activities, but cybersecurity researchers who are approved to join Anthropic’s Cyber Verification Program gain access to AI capabilities in this restricted gray zone.

Also: This new Claude Code Review tool uses AI agents to check your pull requests for bugs – here’s how

Effectively, those able to obtain a security clearance from Anthropic can use Opus 4.7 to perform blocked security activities in the course of doing their job. Disclosure: I am an authorized member of Anthropic’s Cyber Verification Program, so I have access to these capabilities as part of my cyberwarfare, cyberdefense, and counterterrorism work.

Making vulnerabilities actionable

The problem with vulnerability scanning is that it can become a firehose of noise. Every little thing can be flagged, and you can spend hours or days chasing down a bug that is of fairly little consequence instead of repairing a vulnerability that can cause an extinction-level event.

Claude Security is introducing a “multi-stage validation pipeline independently verifies each finding before it reaches an analyst, and every result gets a confidence rating.”

The AI is able to explain each “finding” in detail, including factors like confidence, severity, likely impact, reproduction steps, and recommended fix. This can be enormously helpful, because developers can then prioritize working on those high-confidence, large-impact, severely troubling problems first, without having to waste time on lesser issues.

Also: Why AI is both a curse and a blessing to open-source software – according to developers

From these findings, Claude Security gives defenders the ability to open the code in Claude Code, in context, so they can see and modify the areas needing work right from the finding results.

Anthropic has also added a series of workflow optimizations. It says, “We’ve added scheduled scans for ongoing coverage, the ability to dismiss findings with documented reasons (so future reviewers can trust prior triage decisions), and CSV and Markdown export for integrating findings into existing tracking and audit systems.”

Stay safe out there

Claude Security subscribers can work with technology and security partners. Anthropic specifically pointed out technology partners including CrowdStrike, Palo Alto Networks, SentinelOne, Trend.ai, and Wiz, which are integrating Opus 4.7 into their cybersecurity platforms.

Also: Google bets $32B on AI agent cyber force as security arms race escalates

The company is also working with security partners including Accenture, BCG, Deloitte, Infosys, and PwC, which are deploying Claude Security to help enterprises strengthen their security posture.

Do you see AI vulnerability scanning as more useful for finding dangerous flaws or for helping developers prioritize fixes faster? Let us know in the comments below.


You can follow my day-to-day project updates on social media. Be sure to subscribe to my weekly update newsletter, and follow me on Twitter/X at @DavidGewirtz, on Facebook at Facebook.com/DavidGewirtz, on Instagram at Instagram.com/DavidGewirtz, on Bluesky at @DavidGewirtz.com, and on YouTube at YouTube.com/DavidGewirtzTV.





Source link

Leave a Reply

Subscribe to Our Newsletter

Get our latest articles delivered straight to your inbox. No spam, we promise.

Recent Reviews


The Samsung Keyboard supports glide typing, voice dictation, multiple languages, and deep customization through Good Lock. On paper, it’s a very capable and perfectly functional keyboard. However, it’s only when I started using it that I realized great features don’t necessarily translate to a great user experience. Here’s every problem I faced with the Samsung Keyboard, and why I’m permanently sticking with Gboard as my main Android keyboard.

I have been using Gboard and the Samsung Keyboard on a recently bought Galaxy S24, which I got at a massive discount.

Google’s voice typing doesn’t cut me off mid-sentence

Fewer corrections, fewer cutoffs, faster dictation

I might be a professional writer, but I hate typing—whether it’s on a physical keyboard or a virtual one. I type slower than I think, which I suspect is true for most people. That becomes a problem when I have multiple ideas in my head and need to get them down fast. It’s happened far too often: I start typing one idea and forget the other. Since jacking my brain into a computer isn’t an option (yet), I’ve been leaning more and more on voice typing as the fastest way to capture my thoughts.

Now, both Samsung Keyboard and Gboard support voice typing, but I’ve noticed that Gboard with Google’s voice engine is just better at transcription accuracy. It picks up on accents flawlessly and manages to output the right words. In my experience, it also seems to have a more up-to-date dictionary. When I mention a proper noun—something recently trending like a video game or a movie name—Samsung’s voice typing fails to catch it, but Google nails it.

That said, you can choose Google as your preferred voice typing engine inside Samsung Keyboard, but it’s a buggy experience. I’ve noticed that the transcription gets cut off while I’m in the middle of talking—even when I haven’t taken a long pause. This can be a real problem when I’m transcribing hands-free.

Gboard offers a more accurate glide typing experience

Google accurately maps my swipe gestures to the right words

Voice typing isn’t always possible, especially when you’re in a crowded place and want to be respectful (or secretive). At times like these, I settle for glide (or swipe) typing. It’s generally much faster than tapping on the keyboard—provided the prediction engine maps your gestures to the right word. If it doesn’t, you have to delete that word, draw that gesture again, or worse—type it out manually.

Now, both Samsung Keyboard and Gboard support glide typing, but I’ve noticed Gboard is far more accurate. That said, when I researched this online, I found a 50-50 divide—some people say Gboard is more accurate, others say Samsung is. I do have a theory on why this happens.

Before my Galaxy S24, I used a Pixel 6a, before that a Xiaomi, and before that a Nokia 6.1 Plus. All of my past smartphones came with Gboard by default. I believe Gboard learned my typing patterns over time—what word correlates to what gesture, which corrections I accept, and which ones I reject. After a decade of building up that prediction model, Gboard knows what I mean when my thumb traces a particular shape. Samsung Keyboard, on the other hand, is starting from zero on this Galaxy S24—leading to all the prediction errors. At least that’s my working theory.

There’s also the argument for muscle memory. While glide typing, you need to hit all the correct keycaps for the prediction engine to work. If you’re even off by a slight amount, the prediction model might think you meant to hit “S” instead of “W.” Now, because of my years of typing on Gboard, it’s likely that my muscle memory is optimized for its specific layout and has trouble adapting to Samsung’s.

Swiping vs typing.


Is Swiping Really Faster Than Typing on a Phone Keyboard?

Which typing method reigns supreme?

I mix three languages in one message, and Gboard just gets it

Predictive multilingual typing doesn’t get any better than this

I’m trilingual—I speak English, Hindi, and Bengali. When I’m messaging my friends and family, we’re basically code-mixing—jumping between languages in the same sentence using the Latin alphabet. Now, my friends and I have noticed that Gboard handles code-mixing much more seamlessly than Samsung Keyboard.

If you just have the English dictionary enabled, neither keyboard can guess that you’re trying to transliterate a different language into English. It’ll always try to autocorrect everything, which breaks the flow. The only way to fix this is by downloading a transliteration dictionary like Hinglish (Hindi + English) or Bangla (Latin). Both Samsung Keyboard and Gboard support these dictionaries, but the problem with Samsung Keyboard is that it can only use one dictionary at a time.

Let’s say I’m writing something in Latinized Bangla and suddenly drop a Hindi phrase. Samsung Keyboard will attempt to autocorrect those Hindi words. Gboard is more context-aware. Since my Hinglish keyboard is already installed, I don’t have to manually switch to it. Gboard can detect that I’m using a Hindi word even with the English or Bangla keyboard enabled, and it won’t try to autocorrect what I’m writing. This also works flawlessly with glide typing, which is a huge quality-of-life improvement over Samsung Keyboard.

This isn’t just an India-specific thing either. Code-mixing is how billions of people type every day—Spanglish in the US, Taglish in the Philippines, Franglais across parts of Europe and Africa.

Gboard looks good without me spending an hour on it

I don’t have time for manual customization

Samsung Keyboard is hands down the more customizable option, especially if you combine it with the Keys Cafe module inside Good Lock. You get granular control over almost every aspect of the keyboard—key colors, keycaps, gesture animations, and a whole lot more. While for some users, this is heaven, I just find it too overcomplicated and a massive time sink.

I don’t have the patience to sit and adjust every visual detail of my keyboard. Sure, it gets stale after a while, and you’d want to freshen it up, but I don’t want to spend the better part of an hour tweaking a virtual keyboard. This is where Gboard wins (at least for me) by doing less.

Android 16 brings Material 3 Expressive, which automatically themes your system apps using your wallpaper’s color scheme. With Gboard, all you have to do is change the wallpaper, and the keyboard updates to match—no Good Lock, no manual color picking. It’s a cleaner, more seamless way to keep your phone looking good without putting in the extra legwork.


The keyboard you don’t think about is the one that’s working

I didn’t switch to Gboard because Samsung Keyboard was broken. I switched because Gboard made typing feel effortless. If you’re a Samsung user who’s never tried it, it’s a free download and a five-second switch. You might not go back either.

Pixel 7 with the 8vim keyboard.


I Tried the Weirdest Android Keyboards So You Don’t Have To

Can strange layouts and gestures beat the good old-fashioned QWERTY?



Source link