When Apple Research quietly published The Illusion of Thinking, it expected to stir academic debate. Instead, it triggered a full‑blown brawl over whether today’s large language models (LLMs) genuinely “reason” or merely string words together. The paper’s core claim is stark: chain‑of‑thought prompting appears useful on medium‑tier tasks but collapses when puzzles reach serious complexity. […]