BotBeat
...
← Back

> ▌

NVIDIANVIDIA
PRODUCT LAUNCHNVIDIA2026-03-18

NVIDIA Introduces Kimodo: Large-Scale Kinematic Motion Diffusion Model for Human and Robot Motion Generation

Key Takeaways

  • ▸Kimodo is trained on 700 hours of optical mocap data, substantially larger than previous public datasets, enabling superior motion quality and generalization
  • ▸The model supports multiple control modalities including text prompts, full-body keyframe constraints, end-effector positioning, and path-based navigation
  • ▸A two-stage denoiser architecture minimizes common artifacts like foot skating and floating while maintaining natural motion characteristics
Source:
Hacker Newshttps://research.nvidia.com/labs/sil/projects/kimodo/↗

Summary

NVIDIA has unveiled Kimodo, a kinematic motion diffusion model trained on 700 hours of optical motion capture data that generates high-quality 3D human and robot motions. The model can be controlled through intuitive text prompts and a comprehensive suite of kinematic constraints, including full-body keyframes, sparse joint positions/rotations, 2D waypoints, and dense 2D paths. This represents a significant advancement over previous motion generation systems, which were limited by small-scale public mocap datasets that constrained motion quality, control accuracy, and generalization capabilities.

Kimodo features a carefully designed motion representation and two-stage denoiser architecture that decomposes root and body prediction to minimize motion artifacts while enabling flexible constraint conditioning. Key capabilities include text-to-motion generation, compositional locomotion, full-body constraints, environment interaction through end-effector constraints, and path-based global character translation. The model demonstrates practical applications in robotics, where it can generate humanoid demonstration data more efficiently than traditional teleoperation, with motions exportable to formats compatible with ProtoMotions and MuJoCo for physics-based policy training.

  • The model has direct applications in robotics for generating training data and in entertainment/simulation for character animation
  • NVIDIA provides open-source code, Python API, and a motion authoring demo for practical implementation

Editorial Opinion

Kimodo represents a meaningful step forward in controllable motion generation by addressing a critical bottleneck—the scarcity of large-scale training data. The combination of substantial dataset scale (700 hours of mocap) with a thoughtfully designed architecture that explicitly handles common motion artifacts could make this a valuable tool for roboticists and animation professionals. However, the practical impact will depend on how accessible the model becomes and whether the quality gains justify adoption over existing workflows.

Computer VisionGenerative AIRoboticsEntertainment & MediaResearch

More from NVIDIA

NVIDIANVIDIA
RESEARCH

Nvidia Pivots to Optical Interconnects as Copper Hits Physical Limits, Plans 1,000+ GPU Systems by 2028

2026-04-05
NVIDIANVIDIA
PRODUCT LAUNCH

NVIDIA Introduces Nemotron 3: Open-Source Family of Efficient AI Models with Up to 1M Token Context

2026-04-03
NVIDIANVIDIA
PRODUCT LAUNCH

NVIDIA Claims World's Lowest Cost Per Token for AI Inference

2026-04-03

Comments

Suggested

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
GitHubGitHub
PRODUCT LAUNCH

GitHub Launches Squad: Open Source Multi-Agent AI Framework to Simplify Complex Workflows

2026-04-05
SourceHutSourceHut
INDUSTRY REPORT

SourceHut's Git Service Disrupted by LLM Crawler Botnets

2026-04-05
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us