Diffusion models struggle with meeting various constraints inherent in training data.
A new approach called neural approximate mirror maps (NAMMs) is proposed for handling general, possibly non-convex constraints in diffusion models.
NAMMs learn an approximate mirror map that transforms data into an unconstrained space, enabling reliable generation of valid synthetic data and solving constrained inverse problems.
Experimental results demonstrate the improvement in constraint satisfaction using NAMM-based mirror diffusion models (MDMs) compared to unconstrained diffusion models.