Security Researcher Tests Anthropic's Open-Source Strix AI Penetration Testing Tool, Finds It 'Very Impressive'
Key Takeaways
- ▸Strix, an open-source AI penetration testing tool based on Claude, successfully compromised three Hack The Box machines in 14-40 minutes each, costing $2.66-$8.44 per test
- ▸Model selection is critical: advanced models like GPT-5.3 Codex produced reliable results, while smaller models generated false positives and inconsistent outcomes
- ▸The tool offers significantly easier installation and setup compared to competing agentic frameworks, requiring only basic configuration
Summary
A security researcher has published first impressions of Strix, an open-source autonomous AI penetration testing tool based on Anthropic's Claude models. The tool, which has gained significant traction with over 20,000 GitHub stars, positions itself as an AI agent capable of finding vulnerabilities and creating proof-of-concepts like human security testers. The researcher tested Strix against three retired Hack The Box machines and found that when paired with advanced models like GPT-5.3 Codex, it successfully compromised all targets on the first attempt, completing challenges in 14-40 minutes at costs ranging from $2.66 to $8.44 per machine.
The researcher, writing under the handle 'bearsyankees,' emphasized that model selection is critical to success. Smaller, cheaper models like GPT-5-nano or locally-run open-source alternatives produced unreliable results filled with false positives and random dead ends. However, top-tier coding models demonstrated impressive capabilities, following typical capture-the-flag paths from initial foothold to privilege escalation. The tool's ease of installation stood out compared to other agentic frameworks, requiring only basic environment variable configuration.
While acknowledging the tool's capabilities, the researcher concluded they are 'very impressed' but not yet looking to change careers. The results suggest AI-powered security testing tools are becoming increasingly capable, though questions remain about their readiness to fully replace human penetration testers. The researcher plans to publish additional analysis comparing performance across different AI models and provide practical testing recommendations for security professionals interested in evaluating Strix.
- Despite impressive results, the security researcher remains unconvinced that AI agents will immediately replace human penetration testers
Editorial Opinion
Strix's performance on standardized security challenges demonstrates how rapidly AI agents are advancing in specialized technical domains. The stark performance gap between premium and budget models reveals an emerging truth about agentic AI: capability doesn't scale linearly with model size, and corner-cutting on compute will waste more time and money than it saves. While impressive on retired CTF challenges, the real test will be whether these tools can handle the unpredictability and creative problem-solving required in real-world security assessments.

