Advances in Efficient and High-Quality Generative Modeling

The field of generative modeling is witnessing significant advancements, particularly in the realm of diffusion models. Recent developments are focusing on enhancing the efficiency and quality of image generation processes. Innovations such as multi-stage generation frameworks and the incorporation of visual priors are addressing the limitations of conventional class-guided diffusion models, leading to improvements in texture details and semantic content accuracy. Additionally, there is a growing emphasis on simplifying and stabilizing continuous-time consistency models, which are optimized for fast sampling, by addressing training instability and scaling these models to unprecedented sizes. The introduction of unified frameworks for conditional generation is also notable, offering versatile capabilities with minimal architectural modifications. Notably, the field is also exploring the distillation of discrete diffusion models through dimensional correlations, aiming to speed up generation without compromising quality. These advancements collectively push the boundaries of what is possible in generative modeling, making it more efficient, versatile, and capable of handling complex tasks.

Particularly noteworthy papers include one that introduces a structured identity mapping task to theoretically analyze learning dynamics in diffusion models, and another that proposes a multi-stage generation framework using visual priors to enhance image generation quality.

Sources

Dynamics of Concept Learning and Compositional Generalization

Diffusion Models Need Visual Priors for Image Generation

Distillation of Discrete Diffusion through Dimensional Correlations

Simplifying, Stabilizing and Scaling Continuous-Time Consistency Models

A Simple Approach to Unifying Diffusion-based Conditional Generation

Efficient Diffusion Models: A Comprehensive Survey from Principles to Practices

DDIL: Improved Diffusion Distillation With Imitation Learning

MambaPainter: Neural Stroke-Based Rendering in a Single Step

One Step Diffusion via Shortcut Models

Enhancing Dataset Distillation via Label Inconsistency Elimination and Learning Pattern Refinement

Diffusion Curriculum: Synthetic-to-Real Generative Curriculum Learning via Image-Guided Diffusion

Generation through the lens of learning theory

Built with on top of