This paper from Stanford and Harvard explains why most “agentic AI” systems feel impressive in demos and then completely fall apart in real use.


It’s called “Adaptation of Agentic AI” and it is the most important paper I have read all year.
Right now, everyone is obsessed with building autonomous agents. We give them tools, memory, and a goal, and expect them to do our jobs.
But when deployed in the real world, they hallucinate tool calls. They fail at long-term planning. They break.
Here’s why:
We are trying to cram all the learning into the AI's brain.
When developers try to fix a broken agent, they usually just fine-tune the main model to produce better final answers.
The researchers discovered a fatal flaw in this approach.
If you only reward an AI for getting the final answer right, it gets lazy.
It literally learns to stop using its tools. It tries to guess the answer instead of doing the work. It ignores the calculator and tries to do the math in its head.
To fix this, researchers mapped out a new 4-part framework for how agents should actually learn.
And the biggest takeaway completely flips the current meta.
Instead of constantly retraining the massive, expensive "brain" of the agent, the most reliable systems do the opposite.
They freeze the brain. And they adapt the tools.
They call it Agent-Supervised Tool Adaptation.
#GateSquareAprilPostingChallenge
post-image
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Repost
  • Share
Comment
Add a comment
Add a comment
No comments
  • Pin