According to an in-depth CNBC report, during two closed-door meetings in Silicon Valley this week, multiple AI startup CEOs and engineers said that the current rollout of AI agents at scale faces two major structural problems: “massive token waste” and “extreme chaos between systems.” This on-the-ground record stands in sharp contrast to the optimism expressed by Nvidia CEO Jensen Huang in March when he said AI agents are “the next ChatGPT,” suggesting that the real bottleneck in this space isn’t compute power, but rather decision design, token efficiency, and multi-system integration.
The biggest problem is dumping everything onto the LLM
In the meeting, Meibel CEO Kevin McGrath said: “The biggest problem we’re dealing with right now is the mistaken belief that everything needs to be processed by a large language model—dump all the tokens and all the money into one AI bot, and it will burn through millions of tokens.” He emphasized that when enterprises design agent workflows, they must make clearer judgments about which tasks truly require an LLM, and which can be handled with cheaper rules-based logic or traditional machine learning.
This observation echoes the market reaction after Anthropic’s Claude Enterprise shifted to usage-based pricing—once token consumption is directly tied to cost, the “blindly throw it to an agent” development model immediately reveals its financial pressure. Meibel’s view represents a group of anti-hype engineering pragmatists: the art of agent architecture lies in constraint, not permissiveness.
Chaos in multi-agent collaboration systems that depend on each other
Another key word that keeps appearing in the CNBC report is “chaotic.” When companies run multiple AI agents at the same time—for example, one handling customer service, one handling scheduling, and one handling finance—message passing between agents, state consistency, and error recovery can all affect one another, and any agent behaving abnormally can trigger a cascading spread. Karpathy also mentioned this week that he personally runs workflows with 10–20 agents at the same time, but admitted that code review and the PR process have become a new bottleneck.
The chaos in these multi-agent systems is, at its core, a replay in the LLM era of an old problem in distributed systems: no clear SLA, no transactional boundaries, and no failure-retry semantics. Although Anthropic and OpenAI have rolled out protocol-layer efforts such as MCP and Agent SDK, in real-world enterprise deployments, standardization still lags far behind the growth in the number of agents.
Jensen Huang’s $250k token salary thesis cools off
In March, Nvidia CEO Jensen Huang aggressively promoted the concept of “token salary” at GTC and in subsequent interviews, claiming: “If an engineer earning a $500k annual salary doesn’t consume at least $250k worth of tokens, I’ll feel deeply uneasy.” His logic is that engineers should replace their own low-level actions with AI agents, and the absolute amount of tokens consumed becomes a proxy indicator of productivity. This argument can be found in the full discussion of AI compute demand in Jensen Huang’s latest exclusive interview (Part 1).
But the on-the-ground opinions in the CNBC report show that Silicon Valley’s engineering community is growing increasingly cautious about this claim: the amount of tokens consumed doesn’t equal productivity, and may even be a signal of poorly designed agents. Engineers’ real value still lies in “deciding which tasks are worth calling an agent, how to break down tasks, and how to design error handling”—work that cannot be measured by token consumption alone.
Crypto and AI agents still need time to intersect
For the crypto industry, this week’s trends—AI consuming 80% of global venture capital, and DeFi projects actively integrating self-directed agents—are built on the premise that agent technology has reached a deployable level. But this CNBC report serves as a reminder: even in purely web2 enterprise environments, neither agent token efficiency nor multi-system integration has yet become reliably stable. Putting agents into a 24/7 on-chain environment where assets can be stolen immediately would amplify both technical risk and financial risk. The real starting point for Crypto × AI may have to wait until standardization at the agent framework layer (such as MCP, LangGraph, and Cloudflare Agents) matures.
This article Silicon Valley AI Agent reality: massive token waste, “extreme chaos” in systems integration, Jensen Huang’s “next ChatGPT” prediction still needs verification first appeared on Chain News ABMedia.
Axios exclusive: The U.S. NSA bypassed the Pentagon blacklist to use Anthropic Mythos, and Dario Amodei urgently met with the White House to negotiate
Vitalik confirms a talk in Hong Kong, with Ethereum AI and ZK ecosystem applications as the core focus
AI consumes 80% of global venture capital; Q1 2026 sees a pull of $242 billion: How crypto players should respond to the reallocation of capital
NVIDIA CEO Jensen Huang’s Latest Interview: Can NVIDIA’s Moat Continue to Hold? (Part 1)