Simplifying, Stabilizing and Scaling Continuous-Time Consistency Models
Abstract
Consistency models (CMs) are a powerful class of diffusion-based generative models optimized for fast sampling. Most existing CMs are trained using discretized timesteps, which introduce additional hyperparameters and are prone to discretization errors. While continuous-time formulations can mitigate these issues, their success has been limited by training instability. To address this, we propose a simplified theoretical framework that unifies previous parameterizations of diffusion models and CMs, identifying the root causes of instability. Based on this analysis, we introduce key improvements in diffusion process parameterization, network architecture, and training objectives. These changes enable us to train continuous-time CMs at an unprecedented scale, reaching 1.5B parameters on ImageNet 512x512. Our proposed training algorithm, using only two sampling steps, achieves FID scores of 2.06 on CIFAR-10, 1.48 on ImageNet 64x64, and 1.88 on ImageNet 512x512, narrowing the gap in FID scores with the best existing diffusion models to within 10%.
Community
Simplifying, Stabilizing and Scaling Continuous-Time Consistency Models
arXiv: https://arxiv.org/abs/2410.11081
Waovvv, paper from OpenAI, this is not common these days.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Simple and Fast Distillation of Diffusion Models (2024)
- Target-Driven Distillation: Consistency Distillation with Target Timestep Selection and Decoupled Guidance (2024)
- Simple ReFlow: Improved Techniques for Fast Flow Models (2024)
- OSV: One Step is Enough for High-Quality Image to Video Generation (2024)
- Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper