BotBeat
...
← Back

> ▌

AnthropicAnthropic
RESEARCHAnthropic2026-05-09

AI-Powered Vulnerability Research Platform Discovers 20+ Critical CVEs, Including Remote Linux Kernel Exploits

Key Takeaways

  • ▸Critical remote code execution vulnerabilities discovered in Linux kernel's ksmbd through autonomous LLM-driven research
  • ▸Novel methodology leverages LLMs to find documentation-to-code mismatches, proving effective for large-scale vulnerability discovery
  • ▸Claude and other modern LLMs demonstrated capability to orchestrate complex security research workflows with minimal scaffolding
Source:
Hacker Newshttps://heyitsas.im/posts/drinking-llms/↗

Summary

An autonomous vulnerability-hunting platform powered by large language models has discovered over 20 CVEs in recent months, including two critical remote, unauthenticated out-of-bounds writes in the Linux kernel's ksmbd (CVE-2026-31432 and CVE-2026-31433). The research demonstrates a novel approach to security vulnerability discovery: instead of asking LLMs to directly drive exploit tools, the system leverages them to identify mismatches between code and documentation—a technique inspired by the discovery of a 12-year-old sudo privilege escalation vulnerability.

The platform, built as a custom harness combining multiple LLM models including Claude and Qwen derivatives, achieved significant results by treating vulnerability research as a documentation-to-code comparison problem. The system discovered that modern LLMs are now capable enough to greatly simplify the scaffolding required for context-heavy external tool use, enabling autonomous security research at scale. The ksmbd vulnerabilities found are particularly concerning: attackers can pack multiple file-sharing operations into a single request, causing the kernel to perform insufficient bounds-checking on variable-length metadata, enabling remote exploitation on unpatched systems.

The research validates three key findings: LLMs can effectively identify documentation-code mismatches (answering the original research question), they can discover vulnerabilities more broadly (beyond just mismatch-type bugs), and they show promise for unlocking novel bug classes or enhancing smaller models' hunting capabilities. With dozens of additional findings still under review and publication, this work represents a significant demonstration of AI's emerging role in critical infrastructure security research.

  • 20+ CVEs discovered autonomously, suggesting LLM-powered security research is now viable at production scale

Editorial Opinion

This research represents a watershed moment for AI security tools—moving beyond academic curiosity into practical vulnerability discovery that protects billions of users. The elegant insight of hunting for documentation-code mismatches rather than attempting fully general vulnerability detection shows how AI systems excel when given a focused, well-defined problem space. What's most significant isn't just the CVEs discovered, but the validation that modern LLMs can orchestrate complex, autonomous workflows in safety-critical domains with surprising reliability.

Large Language Models (LLMs)AI AgentsMachine LearningCybersecurityAI Safety & Alignment

More from Anthropic

AnthropicAnthropic
OPEN SOURCE

Anthropic Releases Prempti: Open-Source Guardrails for AI Coding Agents

2026-05-12
AnthropicAnthropic
PRODUCT LAUNCH

Anthropic Unleashes Computer Use: Claude 3.5 Sonnet Now Controls Your Desktop

2026-05-12
AnthropicAnthropic
PARTNERSHIP

SpaceX Backs Anthropic with Massive Data Centre Deal Amidst Musk's OpenAI Legal Battle

2026-05-12

Comments

Suggested

AnthropicAnthropic
OPEN SOURCE

Anthropic Releases Prempti: Open-Source Guardrails for AI Coding Agents

2026-05-12
vlm-runvlm-run
OPEN SOURCE

mm-ctx: Open-Source Multimodal CLI Toolkit Brings Vision Capabilities to AI Agents

2026-05-12
AnthropicAnthropic
PRODUCT LAUNCH

Anthropic Unleashes Computer Use: Claude 3.5 Sonnet Now Controls Your Desktop

2026-05-12
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us