BotBeat
...
← Back

> ▌

AnthropicAnthropic
PRODUCT LAUNCHAnthropic2026-05-05

Anthropic Opens Claude Code ARC-AGI-3 Traces to Public: Transparent AI Reasoning Benchmarking

Key Takeaways

  • ▸Anthropic releases publicly replayable traces of Claude Code solving ARC-AGI-3 reasoning puzzles, enabling transparency in AI agent performance
  • ▸Dashboard shows strong performance with 107 puzzles solved across multiple variants, using different models (Sonnet, Opus) and experimental configurations
  • ▸Full traceability of each run—steps, tokens, timestamps, seeds—allows researchers to reproduce results and understand the agent's reasoning process
Source:
Hacker Newshttps://arc-agi-runs.web.app↗

Summary

Anthropic has launched a public demo dashboard showcasing replayable traces of Claude Code's performance on ARC-AGI-3 (Abstraction and Reasoning Corpus) puzzle games. The interactive dashboard logs detailed run data from multiple variants of Claude Code, including different model sizes (Sonnet, Opus) and experimental configurations, allowing researchers and developers to replay and analyze the agent's problem-solving process step-by-step.

The public demo displays comprehensive metrics across 25 games, with results showing 107 solved puzzles, 16 timeouts, 15 unsolved challenges, and 3 unknown outcomes. Each run is fully traceable—showing individual steps taken, token consumption, variant configuration, completion time, and random seed—providing unprecedented transparency into how AI agents approach abstract reasoning tasks. The interface supports filtering and exploration across multiple experimental conditions, including variants testing whether the agent benefits from previous run context, code-writing capabilities, PNG image support, and stricter filesystem isolation.

This initiative represents a shift toward reproducibility and transparency in AI benchmarking. By making traces publicly accessible, Anthropic enables independent verification of Claude Code's reasoning capabilities while providing researchers with rich data for studying AI problem-solving behavior, failure modes, and the impact of architectural choices on performance.

  • Open access supports research reproducibility and independent verification of AI reasoning capabilities

Editorial Opinion

This move exemplifies the kind of transparency the AI industry needs. By opening replayable traces of Claude Code's attempts at ARC-AGI puzzles, Anthropic provides researchers a window into how agentic AI systems approach abstract reasoning—not just final scores, but the actual reasoning traces. The detailed variant testing (code-writing ablations, context reuse, isolation levels) is particularly valuable for understanding what drives reasoning performance. This is a meaningful step toward making AI development more auditable and reproducible.

Generative AIAI AgentsScience & ResearchOpen Source

More from Anthropic

AnthropicAnthropic
OPEN SOURCE

Anthropic Releases Prempti: Open-Source Guardrails for AI Coding Agents

2026-05-12
AnthropicAnthropic
PRODUCT LAUNCH

Anthropic Unleashes Computer Use: Claude 3.5 Sonnet Now Controls Your Desktop

2026-05-12
AnthropicAnthropic
PARTNERSHIP

SpaceX Backs Anthropic with Massive Data Centre Deal Amidst Musk's OpenAI Legal Battle

2026-05-12

Comments

Suggested

AnthropicAnthropic
OPEN SOURCE

Anthropic Releases Prempti: Open-Source Guardrails for AI Coding Agents

2026-05-12
vlm-runvlm-run
OPEN SOURCE

mm-ctx: Open-Source Multimodal CLI Toolkit Brings Vision Capabilities to AI Agents

2026-05-12
AnthropicAnthropic
PRODUCT LAUNCH

Anthropic Unleashes Computer Use: Claude 3.5 Sonnet Now Controls Your Desktop

2026-05-12
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us