Latest large reasoning models (LRMs) face 'complete accuracy collapse' on highly complex tasks, as per a new paper co-authored by Apple researchers.
LRMs perform well on some problems but struggle on highly complicated ones, indicating a limit to their performance.
Researchers used controllable puzzles to analyze LRMs' reasoning breakdown as difficulty increased, leading to 'complete accuracy collapse' beyond a certain complexity level.
Study reveals LRMs reduce reasoning effort on harder problems, showcasing fundamental limitations in how they handle complex tasks, raising questions about achieving truly general AI.