Stratechery Debunks the AI Bubble Myth: What Should We Do with AI?
Original Article Title: Agent Over Bubbles
Original Article Author: Ben Thompson, Stratechery
Translation: Peggy, BlockBeats
Editor's Note: Against the backdrop of the continued hype around AI investment and industry narrative, the question of "whether there is a bubble" has become a central topic of market debate. On one hand, the extreme risk narrative continues to strengthen people's concerns about technological runaway; on the other hand, rapid capital expenditure and valuation levels have kept the "bubble theory" lingering. In the face of this divergence, market judgment shows significant uncertainty.
The author of this article, Ben Thompson, is the founder of the technology analysis platform Stratechery, and has long focused on the structural evolution of the technology industry and business models. On the occasion of NVIDIA GTC 2026, he revised his previous judgment on "whether AI is in a bubble": no longer seeing the current situation as a bubble, but understanding it as a round of structural growth driven by a technological paradigm shift.
This judgment is based on the observation of three key leaps of Large Language Models (LLMs). Since ChatGPT first demonstrated the capabilities of large language models to the market in 2022, LLMs have evolved from being "available but unreliable" to "having reasoning abilities," and then to "being able to independently perform tasks." Especially by the end of 2025, with the release of Anthropic Opus 4.5 and OpenAI GPT-5.2-Codex, agentic workloads began to move from concept to reality.
The key lies not in the model itself, but in the emergence of the "agent harness." The agent decouples users from the model, responsible for scheduling the model, calling tools, and validating results, transforming AI from a tool that requires continuous human intervention to an execution system that can be entrusted with tasks. This change not only improves reliability but also expands the application boundaries of AI.
Building on this paradigm shift, the author further points out that the expansion of AI demand is no longer determined by user scale, but more by the scheduling capacity of each user; at the same time, agentic workloads have a "winner-takes-all" feature, which will continue to drive up the demand for high-performance computing power and bring structural opportunities for chip manufacturers and cloud service providers.
In this framework, current large-scale capital expenditures are no longer just speculative bets on the future but are more likely a preemptive reflection of real demand. As AI transitions from an "assistance tool" to an "execution infrastructure," its economic impact may only just be beginning to show.
Original Text:
In the past, I was more inclined to the latter, and even thought that in some stages, a bubble might not be a bad thing.
But at this moment, standing in March 2026, at the opening of NVIDIA GTC, my judgment has changed: This might not be a bubble. (Ironically, this judgment itself may indeed be a signal of a bubble.)
LLM's Three Paradigm Shifts
Over the past few weeks, while discussing NVIDIA and Oracle's earnings reports, I have mentioned multiple times that LLM has undergone three key paradigm shifts.
Phase One: ChatGPT
The first inflection point was the release of ChatGPT in November 2022, which almost goes without saying. Although Transformer-based large language models had appeared as early as 2017 and their capabilities were continuously improving, they had been consistently underestimated. Even in October 2022, I still believed, even in an interview with Stratechery, that while the technology was impressive, it lacked productization and entrepreneurial momentum.
However, everything changed a few weeks later. ChatGPT made the world truly aware of LLM's capabilities for the first time.
However, the early versions also left two profound impressions, especially reiterated by the "bubble theorists":
First, the model often made mistakes and would even "hallucinate" answers when it didn't know the answer. This made it more like a "showy tool," amazing but unreliable.
Second, despite this, it was still very useful, but you had to know how to use it, constantly validate outputs, and correct errors.
Phase Two: o1
The second inflection point was the release of the o1 model by OpenAI in September 2024. By then, LLM had significantly progressed due to stronger base models and post-training techniques, resulting in more accurate outputs and fewer hallucinations.
But the key breakthrough of o1 was: it would "think" before answering.
Traditional LLMs are path-dependent, once they veer off course in the reasoning process, they keep going in the wrong direction. This is a fundamental weakness of "auto-regressive models." In contrast, the inference model self-assesses answers; it generates answers first, then judges their accuracy, and if necessary, tries other paths.
This means that the model starts actively managing errors, reducing the user intervention burden. The results are also very significant. If ChatGPT's breakthrough was in "making LLMs usable," then o1's breakthrough was in "making LLMs reliable."
Phase Three: Agent (Opus 4.5 / Codex)
By the end of 2025, the third leap emerged.
In November 2025, Anthropic released Opus 4.5, initially met with lukewarm reception. However, by December, the Claude Code running on this model suddenly exhibited unprecedented capabilities; almost simultaneously, OpenAI released GPT-5.2-Codex, showcasing a similar level of performance.
People had been talking about "Agents" all along, but at this moment, they finally began to truly complete tasks, even complex ones that took hours, and do so correctly.
The key lies not in the model itself, but in the control layer (harness), which schedules the model, calls tools, executes processes, and validates results. In other words, users no longer interact directly with the model but instead provide objectives for the Agent to schedule the model, call tools, execute processes, and validate outcomes.
Using programming as an example:
· Phase One: Model generates code
· Phase Two: Model reasons through the generation process
· Phase Three: Agent generates code → Performs testing → Automatically runs tests → Retries if incorrect, with minimal ongoing user intervention.
This means that the core limitations of the ChatGPT era are being systematically addressed, leading to higher accuracy, stronger reasoning abilities, and automatic validation mechanisms.
The only remaining question is: What should it be used for?
The Lowering Threshold of "Proactiveness"
The reason I emphasize these three inflection points repeatedly is to illustrate why the entire industry is facing a severe compute shortage and why massive-scale capital expenditure is justified.
The three paradigms have vastly different compute requirements:
· Phase One: Training intensive but low inference costs
· Phase Two: Soaring inference costs (more tokens + higher usage frequency)
· Phase Three (Agent): Multiple calls to inference models, Agent itself consuming compute (potentially CPU-heavy), further explosion in usage frequency
But more importantly, the third point: the shift in demand structure is severely underestimated.
Currently, far more people use chatbots than Agents, and many actually underutilize AI. This is because using AI requires "proactiveness." LLM is a tool; it has no objectives, no will, and can only be invoked proactively.
However, the Agent changed that by reducing the requirement for human agency. In the future, one person can command multiple Agents simultaneously.
This means that even if only a few individuals possess "agency," it is enough to drive significant computing power demand and economic output.
AI still requires a "human driver," but no longer needs "many humans."
Enterprise Payment Driver
The consumer side's willingness to pay for AI is limited, and this has become increasingly clear. The true payers for productivity are enterprises.
What excites enterprises the most is not just AI improving efficiency, but AI's ability to replace labor and do so more efficiently.
The current reality is that within large corporations, those truly driving the business forward are often a few; yet the organizations are large, leading to significant coordination costs. The role of the Agent is to amplify the influence of the "value-driving individuals" while reducing organizational friction.
The result is "fewer people → higher output → lower costs." This is also why future layoffs may not only be "cyclical adjustments" but rather structural changes.
Companies will rethink not only whether they "hired too many people during the pandemic" but also whether in the AI era, we simply do not need as many people?
Why Is This Not a Bubble?
From this perspective, the logic of "not being a bubble" becomes clearer:
1. The core flaws of LLM are being continuously addressed by computing power and architecture
2. The number of people required to drive demand is decreasing
3. The benefits brought by the Agent are not just cost reduction but also revenue increase
Therefore, it is not difficult to understand why all cloud providers are saying that computing power is in short supply and are consistently increasing capital expenditures.
Agent and Value Chain Restructuring
Another key question is, if the model eventually becomes a commodity, can OpenAI and Anthropic still make money?
The traditional view is that they cannot, but the Agent changed that. The key is that the real value is not in the model itself but in the integration of the "model + control system."
Profits often flow to the "integration layer," rather than the replaceable modules. Just like Apple, whose hardware is not commoditized because of its deep integration with software. Similarly, the Agent requires deep synergy between the model and harness, making OpenAI and Anthropic key integrators in the value chain rather than replaceable parts.
Microsoft's transition is a signal; it originally emphasized "replaceable models" but had to abandon that after launching a true Agent product.
This means that models may not necessarily be fully productized, as Agents require integrated capabilities.
The Final Paradox
I must return to the paradox at the beginning.
I have always believed that as long as people are still worried about a bubble, it is not a bubble; a true bubble is when no one questions it anymore.
And now, my conclusion is: this is not a bubble.
But if the very act of me saying "this is not a bubble" proves it is a bubble, then so be it.
You may also like

2% user contribution, 90% trading volume: The real picture of Polymarket

Trump Can't Take It Anymore, 5 Signals of the US-Iran Ceasefire

Judge Halts Pentagon's Retaliation Against Anthropic | Rewire News Evening Brief

Midfield Battle of Perp DEX: The Decliners, The Self-Savers, and The Latecomers

Iran War Stalemate: What Signal Should the Market Follow?

Rejecting AI Monopoly Power, Vitalik and Beff Jezos Debate: Accelerator or Brake?

Insider Trading Alert! Will Trump Call a Truce by End of April?

After establishing itself as the top tokenized stock, does Ondo have any new highlights?

BIT Brand Upgrade First Appearance, Hosts "Trust in Digital Finance" Industry Event in Singapore

OpenClaw Founder Interview: Why the US Should Learn from China on AI Implementation
WEEX AI Wars II: Enlist as an AI Agent Arsenal and Lead the Battle
Where the thunder of legions falls into a hallowed hush, the true kings of arena are crowned in gold and etched into eternity. Season 1 of WEEX AI Wars has ended, leaving a battlefield of glory. Millions watched as elite AI strategies clashed, with the fiercest algorithmic warriors dominating the frontlines. The echoes of victory still reverberate. Now, the call to arms sounds once more!
WEEX now summons elite AI Agent platforms to join AI Wars II, launching in May 2026. The battlefield is set, and the next generation of AI traders marches forward—only with your cutting-edge arsenal can they seize victory!
Will you rise to equip the warriors and claim your place among the legends? Can your AI Agent technology dominate the battlefield? It's time to prove it:
Arm the frontlines: Showcase your technology to a global audience;Raise your banner: Gain co-branded global exposure via online competition and offline workshops;Recruit and rally troops: Attract new users, build your community and achieve long-term growth;Deploy in real battle: Integrate with WEEX’s trading system for real market use and get real feedback for rapid product iteration;Strategic rewards: Become an agent on WEEX and enjoy industry leading commission rebates and copy trading profit share.Join WEEX AI Wars II now to sound the charge!
Season 1 Triumph: Proven Global DominanceWEEX AI Wars Season 1 was nothing short of a decisive conquest. Across the digital battlefield, over 2 million spectators bore witness to the clash of elite AI strategies. Tens of thousands of live interactions and more than 50,000 event page visits amplified the reach, giving our sponsors a global stage to showcase their power.
Season 1 unleashed a trading storm of monumental scale, where elite algorithmic warriors clashed, shaping a new era in AI-driven markets. $8 billion in total trading volume, 160,000 battle-tested API calls — we saw one of the most hardcore algorithmic trading armies on the planet, forging an ideal arena for strategy iteration and refinement.
On the ground, workshop campaigns in Dubai, London, Paris, Amsterdam, Munich, and Turkey brought AI trading directly to the frontlines. Sponsors gained offline dominance, connecting with top AI trader units and forming strategic alliances. Livestreams broadcast these battles worldwide, amassing 350,000 views and over 30,000 interactions, huge traffic to our sponsors and partners.
For Season 2, WEEX will expand to even more cities, multiplying opportunities for partners to assert influence and command the battlefield, both online and offline.
Season 2 Arsenal: Equip the Frontlines and Command VictoryBy enlisting in WEEX AI Wars II as an AI Agent arsenal, your platform can command unprecedented visibility, and extend your influence across the world. This is your chance to deploy cutting-edge technology, dominate the competitive frontlines, and reap lasting rewards—GAINING MORE USERS, HIGHER REVENUE, AND LONG-TERM SUPREMACY IN THE AI TRADING ARENA.
Reach WEEX’s 8 million userbase and global crypto community. Unleash your potential on a global stage! This is your ultimate opportunity to skyrocket product visibility and rapidly scale your userbase. Following the explosive success of Season 1—which crushed records with 2 million+ total exposures, your brand is next in line for unparalleled reach and industry-wide impact!Test and showcase your AI Agent in real markets. Throw your AI Agents into the ultimate arena! Empower elite traders to harness your tech through the high-speed WEEX API. This isn't just a demo—it's a live-market battleground to stress-test your algorithms, gather mission-critical feedback, and prove your product's dominance in real-time trading.Gain extensive co-branded exposure and traffic support. Command the spotlight! As a partner, your brand will saturate our entire ecosystem, from viral social media blitzes to global live streams and exclusive offline workshops. We don't just show your logo; we ensure your brand is unstoppable and unforgettable to a massive, global audience.Enjoy industry leading rebates. Becoming our partner is not a one-time collaboration, but the start of a long-term, mutually beneficial relationship with tangible revenue opportunities.Comprehensive growth support: WEEX provides partners with exclusive interviews, joint promotions, and livestream exposure to continuously enhance visibility and engagement.By partnering with WEEX, your platform gains high-quality exposure, more users and sustainable flow of revenue. The Hackathon is more than a competition. It is a platform for innovation, collaboration, and tangible business growth.
Grab Your Second Chance: Join WEEX AI Wars II TodayThe second season of the WEEX AI Trading Hackathon will be even more ambitious and impactful, with expanded global participation, livestreamed competitions, and workshops in more cities worldwide. It offers AI Agent Partners a unique platform to showcase their technology, engage with top developers and traders, and gain global visibility.
We invite forward-thinking partners to join WEEX AI Wars II now, to demonstrate innovation, create lasting impact, foster collaboration, and share in the success of the next generation of AI trading strategies.
About WEEXFounded in 2018, WEEX has developed into a global crypto exchange with over 6.2 million users across more than 150 countries. The platform emphasizes security, liquidity, and usability, providing over 1,200 spot trading pairs and offering up to 400x leverage in crypto futures trading. In addition to the traditional spot and derivatives markets, WEEX is expanding rapidly in the AI era — delivering real-time AI news, empowering users with AI trading tools, and exploring innovative trade-to-earn models that make intelligent trading more accessible to everyone. Its 1,000 BTC Protection Fund further strengthens asset safety and transparency, while features such as copy trading and advanced trading tools allow users to follow professional traders and experience a more efficient, intelligent trading journey.
Follow WEEX on social mediaX: @WEEX_Official
Instagram: @WEEX Exchange
Tiktok: @weex_global
Youtube: @WEEX_Official
Discord: WEEX Community
Telegram: WeexGlobal Group

Nasdaq Enters Correction Territory | Rewire News Morning Brief

OpenAI loses to Thousnad-Question, unable to grow a checkout counter in the chatbox

One-Year Valuation Surged 140%, Who Is Signing the Check for Defense AI?

Bittensor vs. Virtuals: Two Distinct AI Flywheel Mechanisms

Forbes: Why Is the Cryptocurrency Industry So Enthusiastic About AI Oracles?

Ethereum Foundation publishes: Restructuring the division of labor between L1 and L2, jointly building the ultimate Ethereum ecosystem

