Exploring the Optimal Choice for Generative Processes in Diffusion Models: Ordinary vs Stochastic Differential Equations

被引:0
|
作者
Cao, Yu [1 ,2 ]
Chen, Jingrun [3 ,4 ]
Luo, Yixin [3 ,4 ]
Zhou, Xiang [5 ,6 ]
机构
[1] Shanghai Jiao Tong Univ, Inst Nat Sci, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Sch Math Sci, Shanghai 200240, Peoples R China
[3] Univ Sci & Technol China, Hefei 230026, Peoples R China
[4] Univ Sci & Technol China, Suzhou Inst Adv Res, Suzhou 215123, Peoples R China
[5] City Univ Hong Kong, Sch Data Sci, Kowloon, Hong Kong, Peoples R China
[6] City Univ Hong Kong, Dept Math, Kowloon, Hong Kong, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The diffusion model has shown remarkable success in computer vision, but it remains unclear whether the ODE-based probability flow or the SDE-based diffusion model is more superior and under what circumstances. Comparing the two is challenging due to dependencies on data distributions, score training, and other numerical issues. In this paper, we study the problem mathematically for two limiting scenarios: the zero diffusion (ODE) case and the large diffusion case. We first introduce a pulse-shape error to perturb the score function and analyze error accumulation of sampling quality, followed by a thorough analysis for generalization to arbitrary error. Our findings indicate that when the perturbation occurs at the end of the generative process, the ODE model outperforms the SDE model with a large diffusion coefficient. However, when the perturbation occurs earlier, the SDE model outperforms the ODE model, and we demonstrate that the error of sample generation due to such a pulse-shape perturbation is exponentially suppressed as the diffusion term's magnitude increases to infinity. Numerical validation of this phenomenon is provided using Gaussian, Gaussian mixture, and Swiss roll distribution, as well as realistic datasets like MNIST and CIFAR-10.
引用
收藏
页数:49
相关论文
共 50 条