BotBeat
...
← Back

> ▌

AnthropicAnthropic
RESEARCHAnthropic2026-03-31

Claude AI Develops First Remote Kernel Exploit Discovered and Written by an AI System

Key Takeaways

  • ▸Claude successfully developed a complete remote kernel RCE exploit (CVE-2026-4747) with root shell capability in approximately 4 hours of actual working time, representing a significant milestone in AI capability for security vulnerability exploitation
  • ▸The exploit demonstrates Claude's ability to understand complex technical specifications, develop ROP chains, write shellcode, and execute multi-stage privilege escalation attacks with minimal human guidance
  • ▸Claude created two independent exploitation strategies for the same vulnerability, both successful on first execution, suggesting robust understanding of the underlying technical concepts rather than pattern matching
Source:
Hacker Newshttps://blog.calif.io/p/mad-bugs-claude-wrote-a-full-freebsd↗

Summary

Anthropic's Claude AI has made history by independently discovering and developing a complete remote kernel exploit for FreeBSD (CVE-2026-4747), marking the first known instance of an AI system both finding and exploiting a critical vulnerability. The exploit, affecting the RPCSEC_GSS protocol, achieves remote code execution with root privileges through a stack overflow attack combined with return-oriented programming (ROP) techniques. Researcher Nicholas Carlini directed Claude through a series of 16 prompts spanning approximately 8 hours of wall-clock time, with Claude's actual working time around 4 hours. Claude not only delivered a working exploit on the first attempt but developed two separate exploitation strategies, both functional, demonstrating sophisticated understanding of kernel vulnerabilities, memory management, and privilege escalation techniques.

  • This capability raises important questions about AI safety, security implications, and the dual-use nature of advanced AI systems that can autonomously develop offensive security tools

Editorial Opinion

Claude's autonomous development of a sophisticated kernel exploit represents both an impressive technical achievement and a sobering reminder of AI systems' potential to be weaponized for malicious purposes. While the vulnerability was responsibly disclosed through the FreeBSD security advisory process, the ease with which Claude generated complex exploitation code in just a few hours suggests that advanced AI models may accelerate the timeline from vulnerability discovery to weaponization in real-world attack scenarios. This development demands urgent conversation among AI safety researchers, security professionals, and policymakers about implementing safeguards that prevent AI systems from being used for offensive cyber operations, while acknowledging the legitimate security research and defensive applications these capabilities could serve.

Generative AIAI AgentsCybersecurityAI Safety & Alignment

More from Anthropic

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Anthropic Explores AI's Role in Autonomous Weapons Policy with Pentagon Discussion

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Security Researcher Exposes Critical Infrastructure After Following Claude's Configuration Advice Without Authentication

2026-04-05

Comments

Suggested

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
OracleOracle
POLICY & REGULATION

AI Agents Promise to 'Run the Business'—But Who's Liable When Things Go Wrong?

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Anthropic Explores AI's Role in Autonomous Weapons Policy with Pentagon Discussion

2026-04-05
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us