BotBeat
...
← Back

> ▌

NVIDIANVIDIA
RESEARCHNVIDIA2026-03-26

NVIDIA Releases Nemotron-Cascade 2: Compact 30B MoE Model Achieves Gold Medal Performance in IMO and IOI

Key Takeaways

  • ▸Nemotron-Cascade 2 achieves Gold Medal-level IMO and IOI performance with only 3B activated parameters, demonstrating unprecedented intelligence density for open-weight models
  • ▸Expanded Cascade RL and novel multi-domain on-policy distillation techniques enable competitive reasoning performance in a compact MoE architecture
  • ▸Open-weight release of model checkpoints and training data democratizes access to advanced post-training techniques and high-capability reasoning models
Source:
Hacker Newshttps://arxiv.org/abs/2603.19220↗

Summary

NVIDIA has introduced Nemotron-Cascade 2, an open-weight 30B mixture-of-experts (MoE) model with just 3B activated parameters that delivers competitive reasoning and agentic capabilities despite its compact size. The model achieves remarkable performance on prestigious benchmarks, becoming only the second open-weight LLM to earn Gold Medal-level scores in the 2025 International Mathematical Olympiad (IMO), International Olympiad in Informatics (IOI), and ICPC World Finals—accomplishing this with 20x fewer parameters than its nearest competitor. This breakthrough demonstrates exceptional "intelligence density" by combining mathematical and coding reasoning capabilities typically found only in much larger frontier models.

The model's success stems from two key technical advancements over its predecessor, Nemotron-Cascade 1. First, NVIDIA significantly expanded Cascade Reinforcement Learning to cover a broader spectrum of reasoning and agentic tasks. Second, the team introduced multi-domain on-policy distillation from the strongest intermediate teacher models for each domain throughout the training process, allowing efficient recovery from benchmark regressions while maintaining strong performance gains. NVIDIA has released both the model checkpoints and training data to the open-source community, advancing research in efficient post-training methodologies.

Editorial Opinion

Nemotron-Cascade 2 represents a significant milestone in making frontier-level reasoning capabilities accessible through more efficient models. The achievement of IMO/IOI Gold Medals with 20x fewer parameters than competitors challenges assumptions about the compute requirements for advanced reasoning and suggests the post-training methodology may be as important as raw model scale. However, the practical applicability of this 30B MoE model for real-world agentic tasks beyond specialized math and coding domains remains to be seen, and the release's true impact will depend on community adoption and reproducibility.

Large Language Models (LLMs)Generative AIReinforcement LearningAI AgentsOpen Source

More from NVIDIA

NVIDIANVIDIA
RESEARCH

Nvidia Pivots to Optical Interconnects as Copper Hits Physical Limits, Plans 1,000+ GPU Systems by 2028

2026-04-05
NVIDIANVIDIA
PRODUCT LAUNCH

NVIDIA Introduces Nemotron 3: Open-Source Family of Efficient AI Models with Up to 1M Token Context

2026-04-03
NVIDIANVIDIA
PRODUCT LAUNCH

NVIDIA Claims World's Lowest Cost Per Token for AI Inference

2026-04-03

Comments

Suggested

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
OracleOracle
POLICY & REGULATION

AI Agents Promise to 'Run the Business'—But Who's Liable When Things Go Wrong?

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Anthropic Explores AI's Role in Autonomous Weapons Policy with Pentagon Discussion

2026-04-05
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us