Apple researchers found 'fundamental limitations' in cutting-edge AI models, with large reasoning models facing a 'complete accuracy collapse' in highly complex problems.
Standard AI models performed better than LRMs in low-complexity tasks, while both suffered 'complete collapse' with high-complexity tasks.
LRMs reduce reasoning effort as they near performance collapse, leading to concerns raised by Apple researchers and AI experts like Gary Marcus.
The study indicates a 'fundamental scaling limitation' in the thinking capabilities of current reasoning models, with implications for the development of AI towards artificial general intelligence (AGI).