AI Scientist v3 Refactors Research Automation, Scales Experiments from 1 Hour to 24 Hours with Reviewer Agent
Key Takeaways
- ▸AI Scientist v3 replaces ~5,000 lines of orchestration code with Claude's native agentic capabilities, retaining only a literature search skill and instruction file
- ▸The system introduces a Reviewer agent that provides iterative feedback, scaling research experiments from 1-hour to 24-hour timeframes through review-rebuttal-re-experiment cycles
- ▸Over 15 research ideas across 8 domains have been successfully executed, with infrastructure supporting concurrent jobs via Docker isolation and cloud GPU providers
Summary
Developer Alex Li has released AI Scientist v3, a significant architectural overhaul of the automated research system that now leverages Claude's native agentic capabilities rather than hardcoded workflows. The new version replaces approximately 5,000 lines of orchestration code with a simple instruction file and a single literature search skill, allowing Claude to autonomously manage the entire research process from experiment design to paper writing. The system introduces a Reviewer agent that provides iterative feedback, mimicking the peer review process in academic research. Unlike its predecessor v2, which used a rigid 4-stage pipeline with explicit breadth-first search, v3 treats the conversation history as the search tree and allows the AI to orchestrate itself.
The refactored architecture maintains only essential components: a workspace organized like an Overleaf LaTeX template with folders for baselines, experiments, and ablations, plus a 177-line search-papers skill that interfaces with academic databases like Semantic Scholar, OpenAlex, OpenReview, and CrossRef. Everything else—experiment design, statistical analysis, and LaTeX formatting—relies on Claude's built-in knowledge. Ideas can range from fully structured experiment plans to paragraph-length hypotheses, with the system explicitly instructing the agent to treat initial ideas as seeds that evolve through experimentation and reviewer feedback.
The system has successfully executed over 15 distinct research ideas across eight domains including video QA, tool-augmented image generation, prompt injection defense, and tabular machine learning. Most experiments are designed for API-only execution without GPU requirements, though some leverage local RTX 5080 or cloud GPU providers like Modal. The infrastructure supports scaling to many concurrent jobs with per-job Docker isolation, using GitLab for code storage and a web viewer for monitoring. According to Li, the development philosophy centered on "high restraint from 'vibe skills'"—removing unnecessary guidance and trusting frontier models to already possess domain knowledge that earlier versions attempted to encode explicitly.
- The architecture demonstrates that frontier LLMs already possess sufficient knowledge for research tasks like experiment design and LaTeX formatting, requiring minimal skill engineering
Editorial Opinion
AI Scientist v3 represents a fascinating inflection point in AI-assisted research—the realization that orchestrating complex workflows may itself be an unnecessary layer when the underlying model is sufficiently capable. By stripping away thousands of lines of handcrafted pipeline code, Li demonstrates that modern frontier models like Claude can self-organize research processes that previously required explicit state machines. The addition of a Reviewer agent creates a genuinely interesting feedback loop that mirrors how human researchers actually improve their work, though questions remain about whether AI-generated reviews can provide the kind of critical, paradigm-challenging feedback that drives breakthrough science rather than incremental refinement.


