Diffusion models have gained attention in robotics for generating multi-modal distributions of system states and behaviors.
Ensuring precise control over generated outcomes without compromising realism remains a challenge in diffusion models.
A novel framework is proposed to enhance controllability in diffusion models by leveraging multi-modal prior distributions and enforcing strong modal coupling.
The framework achieves superior fidelity, diversity, and controllability in motion prediction and multi-task control experiments, providing a reliable and scalable solution for controllable motion generation in robotics.