UC Berkeley, Stanford University, RAI Institute
SIGGRAPH 2025
*Indicates Joint First Authors
Diffuse-CLoC enables intuitive steering of physics-based character motion through guided diffusion, combining the flexibility of kinematic motion generation with physical realism.
Zero-Shot Generalization to Downstream Tasks
Dynamic Obstacle Avoidance
Character smoothly navigates around dynamic obstacles.
Static Obstacle Navigation
Character moves while avoiding static obstalces
Task-space Control
Desired right end-effector tracking while maintaining whole-body coordination.
Long-horizon Planning
Complex multi-stage tasks executed with a single pre-trained model.
Motion In-betweening
Generates physically plausible transitions between keyframe poses.
Jumping Between Pillars
Can generalize to jumping over increasing pillar heights despite only seeing jumping on flatground.
Waypoint
Can easily guide the character between different locations.
Interactive Joystick Control
Real-time responsive motion adaptation to user input.
Abstract
We present Diffuse-CLoC, a guided diffusion framework for physics-based look-ahead control that enables intuitive, steerable, and physically realistic motion generation. While existing kinematics motion generation with diffusion models offer intuitive steering capabilities with inference-time conditioning, they often fail to produce physically viable motions. In contrast, recent diffusion-based control policies have shown promise in generating physically realizable motion sequences, but the lack of kinematics prediction limits their steerability. Diffuse-CLoC addresses these challenges through a key insight: modeling the joint distribution of states and actions within a single diffusion model makes action generation steerable by conditioning it on the predicted states. This approach allows us to leverage established conditioning techniques from kinematic motion generation while producing physically realistic motions. As a result, we achieve planning capabilities without the need for a high-level planner. Our method handles a diverse set of unseen long-horizon downstream tasks through a single pre-trained model, including static and dynamic obstacle avoidance, motion in-betweening, and task-space control. Experimental results show that our method significantly outperforms the traditional hierarchical framework of high-level motion diffusion and low-level tracking.
Presentation
Baseline Comparisons
Diffuse-CLoC vs. Kinematic Motion Model with Universal RL motion Tracker
Left: Diffuse-CLoC (Ours)
Jumping over an obstacle
Left: Diffuse-CLoC (Ours)
.png)

