The AI Agent Bubble: Why Most Startups Are Building on Shaky Ground
The artificial intelligence conference circuit has become a breeding ground for speculation about which technologies will dominate the next decade. Yet beneath the enthusiasm and venture capital flowing into AI agent startups lies a troubling consensus: many companies are betting their futures on competitive advantages that may not survive the next few years.
Recent observations from industry observers who spent time at major AI conferences reveal a pattern worth examining. Venture capitalists are increasingly focused on a single metric: revenue per engineer. This shift signals something fundamental about how the artificial intelligence economy is evolving—and it’s raising questions about whether the current batch of startups understand what actually creates lasting value.
The Engineering Efficiency Paradox
The venture capital community has settled on a telling measurement for evaluating machine learning-powered companies. Rather than asking “How much revenue does this company generate?” investors now ask “How much revenue per engineer?” The expectation is that this number should climb consistently.
This metric reflects a seismic shift in how software gets built. As large language models from companies like OpenAI and Anthropic become more capable, the engineering labor required to build certain applications has become dramatically cheaper. What once required teams of specialists can now be assembled by small groups working with modern AI research tools.
The flood of booth presentations at recent conferences underscores this reality. Nearly every vendor was hawking solutions to problems that emerged when AI agents started running in production environments. Observability platforms. Governance frameworks. Supervisor agents. Data infrastructure. The pitch was consistent: “Someone needs to manage these autonomous systems.”
But here’s the uncomfortable question that few are asking: which of these solutions will actually matter in 24 months?
The Fragile Moat Problem
Traditional software companies built defensible positions by bundling expensive engineering work and specialized knowledge into products. Customers paid a fixed price and received access to capabilities they couldn’t build themselves. The software vendor captured value while the customer’s usage often remained suboptimal—a dynamic that actually benefited vendor margins.
This entire model is fragmenting. Engineering costs are approaching zero. Domain expertise that once commanded premium prices is increasingly abundant. Pricing is shifting from fixed subscriptions to token-based models, which compress margins because the underlying infrastructure costs (running the large language models themselves) scale directly with usage.
Every vendor on the conference floor was implicitly betting on a new form of competitive advantage to replace the old one. The question is whether any of these bets will actually stick.
Domain Expertise as Defense
The most popular wager is that encoded specialized knowledge creates defensibility. Consider legal AI platforms where the team includes actual attorneys who understand courtroom procedure and regulatory nuance. In the current moment—when AI still feels like magic—this approach works. Customers trust specialists.
The durability is questionable. Prompt architecture, the underlying instructions that guide ChatGPT and similar systems, is fundamentally portable text. The legal expertise supporting it exists in abundance—there are over a million lawyers practicing in the United States alone. As artificial intelligence capabilities mature, the logical endpoint for this category is open marketplaces where prompt designs circulate freely, with the best versions winning through speed of iteration rather than secrecy.
Companies trying to build proprietary moats around prompt engineering will likely lose to open alternatives that improve faster. This mirrors the broader commodification of software engineering itself, where developers increasingly rely on pre-built repositories rather than building from scratch.
Data Infrastructure and the Connectivity Trap
Another popular thesis centers on data infrastructure—platforms that connect AI agents to enterprise databases, messaging systems, and workflow tools. The appeal is obvious: watch an agent explore a database schema and automatically generate interfaces and dashboards. It feels revolutionary.
Strip away the novelty, and what remains is usually a layer of prompt architecture sitting atop a large language model, connected to a data-ingestion component. As data-access standards mature (the Model Context Protocol is already advancing this), and as prompt engineering wisdom gets absorbed into the training of newer models, the proprietary magic evaporates. Companies will face competition from their own customers’ engineering teams building equivalent systems internally, or from open-source alternatives that perform as well or better.
Observability: The Trust Play
Observability platforms—tools that monitor and explain AI agent behavior—occupy a different position. These vendors might survive, but only if they achieve ubiquity and become synonymous with trustworthiness. Think of how Stripe operates: few financial companies seriously consider competitors because trust has already been earned.
The survivors in this category will likely merge with compliance and audit functions rather than remain pure monitoring solutions. They’ll succeed by becoming indispensable for regulatory approval rather than by protecting secret sauce.
The Real Competitive Advantage: Risk Management
If engineering commoditizes, expertise becomes abundant, and data plumbing standardizes, what actually justifies premium pricing? The answer may be counterintuitive: trust and risk transfer.
In regulated industries, the question a bank must answer isn’t “Can we build this ourselves?” (increasingly, yes). The question is “Who absorbs the risk if this fails?” A company with certified security credentials, a named CEO willing to testify before Congress, legal teams available for emergency calls, and insurance indemnification that covers regulatory violations provides something genuinely defensible.
This reframes the entire competitive landscape. Rather than software companies competing on features, the future might involve financial engineering: regulated insurance companies pricing the risk of AI agent failure in compliance-sensitive sectors. The commodity layer—consisting of large language models, standardized data access protocols, and open-source prompt architectures—sits beneath. The insurance wrapper sits above.
The middle layer, where most current startups are concentrated, faces the greatest pressure. They’re building products on top of commodity infrastructure, betting on moats that look increasingly permeable.
Conclusion: Preparing for the Inevitable Shift
The entrepreneurs and investors observing this landscape should be asking whether their business sits in a sustainable position. Those building pure software solutions without trust-based defensibility should worry about margin compression. Those betting entirely on proprietary knowledge in fields where expertise is abundant should consider how quickly that advantage might erode.
The companies most likely to thrive will be those that understand this transition early and position themselves accordingly—whether that means building commodity substrates, becoming indispensable through trust, or finding ways to become essential to regulated industries managing AI-related risk.
The artificial intelligence revolution is accelerating, but not all participants in this revolution will profit equally from the transition ahead.
Frequently Asked Questions
What metric are venture capitalists now using to evaluate AI companies?
Venture capitalists are increasingly focused on revenue per engineer as the key metric for evaluating artificial intelligence startups. This shift reflects how dramatically cheaper engineering labor has become as large language models from companies like OpenAI and Anthropic mature. The expectation is that this metric should increase consistently over time, indicating efficient scaling.
Why is domain expertise no longer a defensible competitive advantage?
While domain expertise creates value in the short term, prompt architecture—the underlying instructions that guide AI systems—is fundamentally portable text. Specialized knowledge exists in abundance across most fields, and as machine learning capabilities mature, the wisdom gets absorbed into the training of newer models. Open-source alternatives will eventually outcompete closed, proprietary approaches that rely solely on expertise encoding.
What type of AI startup is most likely to succeed long-term?
Companies positioned in regulated industries that help organizations manage risk and ensure compliance are most likely to succeed. Rather than competing on pure software features, sustainable businesses will provide trust-based services—including security certifications, legal accountability, and insurance indemnification. These address the core concern of regulated industries: who bears responsibility if an AI system fails?





