This paper introduces Diffusion Policy, a new way of generating robot behavior by representing a robot's visuomotor policy as a conditional denoising diffusion process. We benchmark Diffusion Policy across 15 different tasks from 4 different robot manipulation benchmarks and find that it consistently outperforms existing state-of-the-art robot learning methods with an average improvement of 46.9%. Diffusion Policy learns the gradient of the action-distribution score function and iteratively optimizes with respect to this gradient field during inference via a series of stochastic Langevin dynamics steps. We find that the diffusion formulation yields powerful advantages when used for robot policies, including gracefully handling multimodal action distributions, being suitable for high-dimensional action spaces, and exhibiting impressive training stability. To fully unlock the potential of diffusion models for visuomotor policy learning on physical robots, this paper presents a set of key technical contributions including the incorporation of receding horizon control, visual conditioning, and the time-series diffusion transformer. We hope this work will help motivate a new generation of policy learning techniques that are able to leverage the powerful generative modeling capabilities of diffusion models. Code, data, and training details are available (diffusion-policy.cs.columbia.edu).
机构:
Mississippi State Univ, Dept Polit Sci & Publ Adm, Publ Management & Policy, Mississippi State, MS 39762 USAMississippi State Univ, Dept Polit Sci & Publ Adm, Publ Management & Policy, Mississippi State, MS 39762 USA
Fay, Daniel L.
Wenger, Jeffrey B.
论文数: 0引用数: 0
h-index: 0
机构:
RAND Corp, Santa Monica, CA USA
Amer Univ, Washington, DC 20016 USAMississippi State Univ, Dept Polit Sci & Publ Adm, Publ Management & Policy, Mississippi State, MS 39762 USA