BotBeat
...
← Back

> ▌

AnthropicAnthropic
POLICY & REGULATIONAnthropic2026-04-16

Anthropic's Tightened Cyber Policy Blocks Authorized Bug Bounty Research, Frustrating Security Researchers

Key Takeaways

  • ▸Anthropic's cyber usage filters in Opus 4.6/4.7 are blocking authorized bug bounty work retroactively, disrupting active researcher workflows
  • ▸The content classifier does not appear to consider authorization context or program scope provided within the conversation, leading to false-positive blocks
  • ▸The verification exception process favors researchers with public CVE disclosures and conference visibility, excluding early-career researchers with paid bounty history who could benefit most from Claude's capabilities
Source:
Hacker Newshttps://news.ycombinator.com/item?id=47796041↗

Summary

Anthropic's updated cyber usage policies in Claude Opus 4.6 and 4.7 are blocking security researchers from conducting authorized vulnerability research and bug bounty work, even when full program authorization and scope details are provided in the model's context. The changes, implemented retroactively, prevent researchers from drafting analysis, proof-of-concept refinements, and other tasks explicitly within authorized bug bounty program scope. The model itself acknowledged the authorization in one instance before being blocked by an API-level filter that apparently does not account for authorization context.

The company's remediation path requires researchers to enroll in a "Cyber Verification Program," but the de facto requirements—including public CVEs, conference talks, or established public track records—appear to exclude early-career security researchers who lack public disclosures despite having legitimate paid vulnerability disclosures on major platforms like HackerOne, Immunefi, and Bugcrowd. Affected researchers report losing productivity mid-session when work is unexpectedly blocked and express frustration that the population most likely to benefit from AI-assisted security research is being excluded from the tool.

  • Researchers are requesting better context awareness in policy enforcement, lower-friction verification pathways, and transparency about which tasks are restricted

Editorial Opinion

While Anthropic's intention to prevent misuse of AI for malicious cyber activities is understandable, the current implementation appears to conflate legitimate security research with offensive activity in ways that harm the very researchers most likely to use AI responsibly. A policy framework that ignores explicit authorization context and excludes researchers based on public visibility rather than demonstrated responsible conduct may ultimately push legitimate researchers toward competitors or less safety-conscious platforms.

CybersecurityEthics & BiasAI Safety & AlignmentPolicy & Regulation

More from Anthropic

AnthropicAnthropic
PARTNERSHIP

White House Pushes US Agencies to Adopt Anthropic's AI Technology

2026-04-17
AnthropicAnthropic
RESEARCH

AI Safety Convergence: Three Major Players Deploy Agent Governance Systems Within Weeks

2026-04-17
AnthropicAnthropic
PRODUCT LAUNCH

Finance Leaders Sound Alarm as Anthropic's Claude Mythos Expands to UK Banks

2026-04-17

Comments

Suggested

AnthropicAnthropic
PARTNERSHIP

White House Pushes US Agencies to Adopt Anthropic's AI Technology

2026-04-17
AnthropicAnthropic
RESEARCH

AI Safety Convergence: Three Major Players Deploy Agent Governance Systems Within Weeks

2026-04-17
OpenAIOpenAI
RESEARCH

When Should AI Step Aside?: Teaching Agents When Humans Want to Intervene

2026-04-17
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us