The goal of AI agents isn't to become autonomous, so i am not sure what you are not understanding here
to continue, i think the biggest misconception and false marketing is around so call Large Reasoning Models. They don't reason, and are architectural the same as an LLM, and have mostly the same training methodology.
LLMs are nothing more than non-linear lossy data compression and approximate pattern matching to predict probabilities of tokens following a matched sequence. LRMs are exactly the same, but have additional data sets with intermediate tokens, and extended RL with the old ideas of chain-of-thought and multi-shot prompting. During inference LRMs produce additional intermediate tokens, but nothing stops you doing that with an LLM and CoT process.
This is shown in the recent paper by Apple showing neither LLMs nor LRMs have any reasoning capabilities, and so there are many problems that current models will be useless at.
It is ironic because AI has a long history of reasoning and planning using symbolic processing. I don't think we will see a big threat to jobs like software engineering until we have models that properly combine symbolic and sub-symbolic processing to apply actual reasoning to the language comprehension. Conceptually this doesn't seem hard but i remember this problem being discussed 25 years when i was studying AI (and alcohol ) as a wee nipper at undergrad level.