Consistent Diffusion Models: Mitigating Sampling Drift by Learning to be Consistent

Giannis Daras · Yuval Dagan · Alex Dimakis · Constantinos Daskalakis

Great Hall & Hall B1+B2 (level 1) #2000
[ ]
Wed 13 Dec 8:45 a.m. PST — 10:45 a.m. PST


Imperfect score-matching leads to a shift between the training and the sampling distribution of diffusion models. Due to the recursive nature of the generation process, errors in previous steps yield sampling iterates that drift away from the training distribution. However, the standard training objective via Denoising Score Matching (DSM) is only designed to optimize over non-drifted data. To train on drifted data, we propose to enforce a \emph{Consistency} property (CP) which states that predictions of the model on its owngenerated data are consistent across time. Theoretically, we show that the differential equation that describes CP together with the one that describes a conservative vector field, have a unique solution given some initial condition. Consequently, if the score is learned well on non-drifted points via DSM (enforcing the true initial condition) then enforcing CP on drifted points propagates true score values. Empirically, we show that enforcing CP improves the generation quality for conditional and unconditional generation on CIFAR-10, and in AFHQ and FFHQ. We open-source our code and models:

Chat is not available.