Questioning DeepSeek-R1 and Claude Thinking fundamentally do not reason! Did Apple's controversial paper fail?

Wallstreetcn
2025.06.09 05:41
portai
I'm PortAI, I can summarize articles.

A paper from the Apple team questions the reasoning capabilities of current AI inference models (such as DeepSeek-R1 and Claude 3.7 Sonnet), arguing that these models are actually just good at memorizing patterns rather than true reasoning. The research shows that although these models have acquired complex self-reflection mechanisms through reinforcement learning, their performance collapses when faced with highly complex problems. Apple's research uses a controlled puzzle environment, revealing that standard LLMs are more efficient on simple problems, while both perform poorly on complex issues