A new study introduces Dualformer, a Transformer model that can operate in two reasoning modes, fast and slow, by training on randomized reasoning traces.
Dualformer outperforms baselines in terms of performance and computational efficiency across all modes, achieving a high optimal rate on maze tasks and producing more diverse reasoning traces.
The model can be configured to execute in either fast or slow mode, or automatically decide which mode to engage at inference time.
Dualformer's capabilities extend beyond task-specific models, showing improved performance in math reasoning problems through Large Language Models fine-tuning.