Reasoning-enhanced large language models (LLMs) generate intermediate reasoning steps prior to generating final answers, excelling in complex problem-solving.
Thinking Intervention is a novel paradigm designed to guide the internal reasoning processes of LLMs by strategically inserting or revising specific thinking tokens.
Comprehensive evaluations show that Thinking Intervention outperforms baseline prompting approaches, achieving significant improvements in instruction following, instruction hierarchy, and safety alignment tasks.
The research on Thinking Intervention offers a promising new avenue for controlling reasoning LLMs.