Recent research introduces EB-Sampler, a method for efficient sampling from masked diffusion models (MDMs) used for language modeling.
EB-Sampler utilizes an Entropy Bounded unmasking procedure to predict multiple unknown tokens with a single function evaluation with predefined error tolerance.
The new sampling method accelerates sampling from MDMs by 2-3x on standard coding and math reasoning benchmarks, with no loss in performance.
EB-Sampler also shows effectiveness in smaller reasoning tasks like maze navigation and Sudoku, tasks where autoregressive models often struggle.