Reinforcement Learning Journal, vol. TBD, 2025, pp. TBD.
Presented at the Reinforcement Learning Conference (RLC), Edmonton, Alberta, Canada, August 5–9, 2025.
This work introduces Disentangled Diffusion Policy (DisDP), an Imitation learning method that enhances robustness by integrating multi-view disentanglement into diffusion-based policies. Robots operating in real-world environments rely on multiple sensory inputs to effectively interact with their surroundings. However, sensors are susceptible to noise, calibration errors, failures, and environmental perturbations. Existing Imitation Learning methods struggle to generalize under such conditions, as they typically assume consistent, noise-free inputs. Disentangled Diffusion Policy (DisDP) addresses this limitation by structuring sensory inputs into shared and private representations, preserving task-relevant global features while retaining distinct details from individual sensors. This structured representation improves resilience against sensor dropouts and perturbations. Evaluations on The RoboColosseum and Libero benchmarks demonstrate that DisDP achieves performance on par with baseline methods while exhibiting greater robustness to sensor variations.
Pankhuri Vanjani, Paul Mattes, Xiaogang Jia, Vedant Dave, and Rudolf Lioutikov. "DisDP: Robust Imitation Learning via Disentangled Diffusion Policies." Reinforcement Learning Journal, vol. TBD, 2025, pp. TBD.
BibTeX:@article{vanjani2025disdp,
title={{DisDP}: {R}obust Imitation Learning via Disentangled Diffusion Policies},
author={Vanjani, Pankhuri and Mattes, Paul and Jia, Xiaogang and Dave, Vedant and Lioutikov, Rudolf},
journal={Reinforcement Learning Journal},
year={2025}
}