Also, they exhibit a counter-intuitive scaling limit: their reasoning energy improves with issue complexity nearly a point, then declines Inspite of possessing an enough token spending plan. By evaluating LRMs with their typical LLM counterparts less than equal inference compute, we identify 3 performance regimes: (1) very low-complexity tasks https://www.youtube.com/watch?v=snr3is5MTiU