Task Time horizons are improving exponentially with doubling times around 4 months per METR. At what timescale would you accept that they "can be strategic"? Theres little reason to think they wont be at multi week or month time horizons very soon. Do you need to be strategic to complete multi month tasks?
>Can an LLM give you an upfront estimate that a task will take multiple months?
>Can it decide intelligently what it would have to change if you said "do what you can to have it ready in half the time?"
Do you think ChatGPT 5.2 Pro can't estimate how long a task might take? Do you think that estimate would necessarily be worse than the estimates, which are notoriously poor, coming from human engineers?
But you can still answer my question. When an LLM can complete a task that takes a person N months or years, is it capable of being strategic?
IMO the onus is to prove that they can be strategic. Otherwise you're asking me to prove a negative.