Efficient Fine-Tuning and Concept Suppression for Pruned Diffusion Models
Diffusion generative models have advanced significantly, but they have become larger and more complex, creating computational challenges in resource-constrained scenarios.
Pruning and knowledge distillation can reduce computational demands while preserving generation quality, but they can also propagate undesirable behaviors.
A new bilevel optimization framework is proposed to consolidate fine-tuning and unlearning processes, selectively suppressing the generation of unwanted content in diffusion models.