Also, they exhibit a counter-intuitive scaling limit: their reasoning exertion raises with challenge complexity as much as a degree, then declines Regardless of acquiring an ample token budget. By comparing LRMs with their typical LLM counterparts under equivalent inference compute, we detect three general performance regimes: (1) very low-complexity https://www.youtube.com/watch?v=snr3is5MTiU