Diffusion Policies for Visuomotor Policy Learning: A Paradigm Shift in Robotics
Robotic manipulation has long been challenged by the need for robust visuomotor policies that can generalize across diverse objects and interaction dynamics. Traditional approaches often fall short due to their reliance on direct observation-to-action mappings or their tendency to compress perceptual inputs into overly simplistic features. A new paradigm is emerging, leveraging the power of diffusion models to overcome these limitations and unlock new possibilities for robot learning. This article explores the concept of diffusion policies, focusing on two innovative frameworks: Diffusion Policy and 3D Flow Diffusion Policy (3D FDP).
The Limitations of Traditional Behavioral Cloning
Behavioral cloning, a dominant method in robot learning from demonstration, treats the problem as supervised learning. The robot observes a state (e.g., an image) and attempts to predict the corresponding action, typically by minimizing a regression loss. While straightforward, this approach struggles with the inherent complexities of real-world scenarios:
- Messy Human Demonstrations: Human demonstrations are rarely perfect. There are often multiple ways to accomplish the same task, actions can be ambiguous, and tasks may require long, coordinated sequences of movements.
- Averaging Problem: Traditional behavioral cloning tends to average over these possibilities, resulting in actions that are indecisive and suboptimal.
- Lack of Temporal Consistency: Generating consistent, smooth action sequences is difficult, often leading to jittery or erratic robot behavior.
Diffusion Models: A Generative Approach to Robot Control
Diffusion models, which have revolutionized image generation, offer a compelling alternative. Instead of directly predicting the next action, a diffusion policy starts with random noise and iteratively refines it into a plausible action (or action sequence) conditioned on the current observations.
The core idea is to learn the gradient of the action distribution, known as the "score function." This function indicates, for any given action and observation, the direction in which the action should be adjusted to increase its likelihood under the distribution of demonstrated behaviors.
Stochastic Langevin Dynamics: Sampling from the Action Distribution
To determine the appropriate action, diffusion policies employ stochastic Langevin dynamics. This process begins with a random action (sampled from Gaussian noise) and uses the score function to guide the action towards probable behaviors. A small amount of noise is added at each step to encourage exploration and prevent the action from getting trapped in suboptimal solutions.
Read also: Revolutionizing Generative Modeling
This approach excels at handling multimodal action distributions, where multiple equally valid actions exist. Traditional policies tend to average these options, producing ambiguous actions. In contrast, diffusion policies can commit to one mode at a time, sampling from the full range of demonstrated behaviors without averaging them away.
Predicting Action Sequences for Temporal Consistency
Diffusion Policy makes a big jump by predicting whole sequences of future actions instead of simply the next step. This is very important for real-world tasks, where actions need to be consistent over time and able to handle pauses or unexpected events. The policy makes sure that the robot’s behavior is smooth and consistent across time by making an entire sequence at once.
For instance, if a demonstration has a pause or a time when the person is stationary, sequence prediction helps the policy figure out when to stay still and when to move, which lowers the chance of getting stuck or making shaky movements. Receding horizon control is a method that lets the robot change its plans by making a new action sequence based on the most recent observations. This is how the robot can respond to new information.
Advantages of Diffusion Policies
Diffusion policies offer several key advantages over traditional behavioral cloning:
- Handling Multimodal Demonstrations: They naturally handle multimodal and ambiguous demonstrations, committing to one valid behavior at a time.
- Scalability to High-Dimensional Action Spaces: They scale gracefully to high-dimensional action spaces, enabling the robot to plan and execute complex, coordinated movements.
- Temporal Consistency: They produce temporally consistent and robust action sequences, leading to smoother and more reliable robot performance.
- Training Stability: Training is more stable than energy-based models, since it avoids the pitfalls of negative sampling and intractable normalization constants.
Empirical results support these claims. Diffusion Policy has consistently outperformed state-of-the-art methods across diverse robot manipulation tasks, demonstrating robustness to latency, perturbations, and the variability of human demonstration data. Diffusion Policy outperforms prior state-of-the-art on 12 tasks across 4 benchmarks with an average success-rate improvement of 46.9%. First row: Average of end-states for each method.
Read also: Debate over Keller ISD ID Rules
3D Flow Diffusion Policy: Leveraging Scene-Level Motion Cues
While diffusion policies offer a powerful framework for visuomotor policy learning, further enhancements can be achieved by incorporating structured intermediate representations. 3D Flow Diffusion Policy (3D FDP) leverages scene-level 3D flow to capture fine-grained local motion cues, addressing a critical limitation of many existing approaches.
Most existing approaches rely on direct observation-to-action mappings or compress perceptual inputs into global or object-centric features, which often overlook localized motion cues critical for precise and contact-rich manipulation. 3D FDP predicts the temporal trajectories of sampled query points and conditions action generation on these interaction-aware flows, implemented jointly within a unified diffusion architecture. This design grounds manipulation in localized dynamics while enabling the policy to reason about broader scene-level consequences of actions.
3D Flow as a Structural Prior
3D FDP utilizes 3D flow as a structured intermediate representation. This allows the policy to:
- Capture Fine-Grained Local Motion Cues: By predicting the temporal trajectories of sampled query points, the policy gains a detailed understanding of local motion.
- Ground Manipulation in Localized Dynamics: Conditioning action generation on interaction-aware flows grounds manipulation in localized dynamics.
- Reason About Scene-Level Consequences: The policy can reason about the broader scene-level consequences of its actions.
Performance and Validation
Extensive experiments on the MetaWorld benchmark have demonstrated the effectiveness of 3D FDP. The framework achieves state-of-the-art performance across 50 tasks, particularly excelling on medium and hard settings. Beyond simulation, 3D FDP has been validated on real-robot tasks, consistently outperforming prior baselines in contact-rich and non-prehensile scenarios.
3D Diffusion Policy (DP3): Incorporating 3D Visual Representations
Imitation learning provides an efficient way to teach robots dexterous skills; however, learning complex skills robustly and generalizablely usually consumes large amounts of human demonstrations. To tackle this challenging problem, we present 3D Diffusion Policy (DP3), a novel visual imitation learning approach that incorporates the power of 3D visual representations into diffusion policies, a class of conditional action generative models. The core design of DP3 is the utilization of a compact 3D visual representation, extracted from sparse point clouds with an efficient point encoder. In our experiments involving 72 simulation tasks, DP3 successfully handles most tasks with just 10 demonstrations and surpasses baselines with a 24.2% relative improvement. In 4 real robot tasks, DP3 demonstrates precise control with a high success rate of 85%, given only 40 demonstrations of each task, and shows excellent generalization abilities in diverse aspects, including space, viewpoint, appearance, and instance. Interestingly, in real robot experiments, DP3 rarely violates safety requirements, in contrast to baseline methods which frequently do, necessitating human intervention. Our extensive evaluation highlights the critical importance of 3D representations in real-world robot learning.
Read also: Key metrics of JHEPM
Reproducing Simulation Benchmark Results
To reproduce simulation benchmark results, the following steps can be taken:
- Install the conda environment on a Linux machine with an Nvidia GPU.
- Run the training script:
python train.py --config-dir=.This will create a directory in the formatdata/outputs/yyyy.mm.dd/hh.mm.ss_<method_name>_<task_name>where configurations, logs, and checkpoints are written. - Launch a local Ray cluster. For large-scale experiments, consider setting up an AWS cluster with autoscaling.
- Run the Ray training script:
python ray_train_multirun.py --config-dir=. - Monitor metrics aggregated from all training runs on the Wandb project
diffusion_policy_metrics.
Key Technical Contributions
To fully unlock the potential of diffusion models for visuomotor policy learning on physical robots, several key technical contributions have been made, including:
- Incorporation of receding horizon control.
- Visual conditioning.
- The time-series diffusion transformer.
tags: #diffusion #policy #visuomotor #policy #learning

