Metropolis-adjusted Langevin correctors using score-based acceptance probabilities, including an exact Bernoulli factory method and a Simpson's rule approximation, reduce sampling bias in diffusion models and improve FID scores.
Score-based generative modeling through stochastic differential equations
4 Pith papers cite this work. Polarity classification is still indexing.
verdicts
UNVERDICTED 4representative citing papers
PGID restores watermark detection in diffusion models by using progressive inversion-denoising cycles to correct latents displaced by removal or forgery attacks.
CrystalREPA closes the representation gap between crystal generators and universal MLIPs via contrastive alignment, yielding more stable and valid generated crystals while revealing that MLIP teacher quality is better predicted by representation distinguishability than by leaderboard accuracy.
Aligning noisy hidden states in diffusion transformers to clean features from pretrained visual encoders speeds up training over 17x and reaches FID 1.42.
citing papers explorer
-
Metropolis-Adjusted Diffusion Models
Metropolis-adjusted Langevin correctors using score-based acceptance probabilities, including an exact Bernoulli factory method and a Simpson's rule approximation, reduce sampling bias in diffusion models and improve FID scores.
-
PGID: Progressive Guided Inversion and Denoising for Robust Watermark Detection
PGID restores watermark detection in diffusion models by using progressive inversion-denoising cycles to correct latents displaced by removal or forgery attacks.
-
CrystalREPA: Transferring Physical Priors from Universal MLIPs to Crystal Generative Models
CrystalREPA closes the representation gap between crystal generators and universal MLIPs via contrastive alignment, yielding more stable and valid generated crystals while revealing that MLIP teacher quality is better predicted by representation distinguishability than by leaderboard accuracy.
-
Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
Aligning noisy hidden states in diffusion transformers to clean features from pretrained visual encoders speeds up training over 17x and reaches FID 1.42.