B-DENSE: Branching For Dense Ensemble Network Learning
New technique uses K-fold expanded channels to preserve structural information lost in standard distillation.
Researchers Cherish Puniani, Tushar Kumar, and team propose B-DENSE, a novel distillation framework for diffusion models. It modifies student architectures to output K-fold expanded channels, with each subset aligning to a specific intermediate step in the teacher's trajectory. This dense multi-branch training reduces discretization errors, allowing the student model to learn solution space navigation earlier, resulting in superior image generation quality compared to baseline distillation methods.
Why It Matters
Enables faster, higher-quality AI image generation by preserving critical structural information during model acceleration.