Why Building AI Agents Is Mostly a Waste of Time | by Shenggang Li | Data Science Collective | Jan, 2026 | Medium
Sign up
Sign in
Sign up
Sign in
Data Science Collective
Advice, insights, and ideas from the Medium data science community
Member-only story
Why Building AI Agents Is Mostly a Waste of Time
The Structural, Mathematical, and Economic Limits of RAG Pipelines
--
130
Share
Over the past two years, a peculiar belief has quietly taken hold in the AI community: that the future of intelligence lies in building agents. Everywhere you look, engineers are creating RAG pipelines, chaining tools together, wrapping large language models with orchestration frameworks, and calling the result “AI systems.”
It looks impressive. It feels productive. It satisfies the engineering instinct to build.
But here is the uncomfortable truth: most AI agents are not progress. They are decoration.
They do not meaningfully extend intelligence. They do not create new economic value. And they rarely survive the next generation of base models. In most cases, building AI agents is not a step toward the future of intelligence — it is a temporary distraction from understanding what intelligence actually is.
This is not a technical argument. It is a structural one.
A Simple Mathematical Reality
At its core, a large language model is a function approximator:
Everything an LLM produces is a transformation of input signals and its learned internal representation…
--
--
130
Published in Data Science Collective
Advice, insights, and ideas from the Medium data science community
Written by Shenggang Li
Responses (130)
Help
Status
About
Careers
Press
Blog
Privacy
Rules
Terms
Text to speech