US Enterprises to Triple AI Infrastructure Budgets by 2028

Apr 14, 2026
Interview
US Enterprises to Triple AI Infrastructure Budgets by 2028

As a preeminent expert in data protection and privacy governance, Vernon Yai has spent years navigating the complex intersection of sensitive information and high-performance computing. With enterprises now facing a seismic shift in how they build and fund their digital futures, Yai’s perspective on the physical and strategic foundations of technology is more critical than ever. As organizations transition from cautious experimentation to massive, permanent deployments, the very architecture of the modern business is being rewritten.

In this discussion, we explore the financial and operational restructuring required to support the next generation of automation. We delve into the strategic shift from cloud-reliance to hardware ownership, the challenges of managing dozens of simultaneous AI pilots, and the rising importance of “neocloud” providers. Yai also provides a detailed look at how leadership roles must evolve to bridge the gap between multi-million dollar infrastructure investments and long-term business value.

AI infrastructure budgets are projected to triple or even quadruple by 2028. How should organizations reallocate funds from traditional IT modernization, and what specific metrics must CIOs present to justify these sustained investments?

The shift we are seeing is a move away from one-off, “set and forget” modernization projects toward a model of sustained, year-over-year capital intensity. To fund this, organizations must ruthlessly audit their legacy maintenance budgets and redirect those “keep the lights on” funds toward high-growth automation. CIOs need to move beyond simple uptime metrics and start presenting “value-per-token” and “compute-to-revenue” ratios to their boards. My prioritization process begins with a rigorous “technical debt” audit to see what can be decommissioned, followed by a pilot-to-production roadmap that focuses on the 30% of projects with the highest immediate ROI. Finally, we establish a dedicated “AI Fund” that is shielded from standard IT budget cuts to ensure that multi-year infrastructure builds aren’t stalled by short-term fiscal fluctuations.

Many enterprises are shifting toward a hybrid model that blends public cloud services with owned physical hardware. What are the primary trade-offs when moving from an Opex-heavy cloud model to Capex-focused infrastructure?

The primary trade-off is the loss of immediate flexibility in exchange for long-term cost sovereignty and performance guarantees. While the public cloud offers the “adrenaline shot” of quick scaling, the sheer volume of AI workloads is making “renting” compute prohibitively expensive for large-scale operations. By owning hardware, a company can scale at its own internal rhythm, free from the crowded resource marketplace where everyone is competing for the same limited GPU clusters. There is a palpable sense of security that comes from knowing your mission-critical AI workloads aren’t subject to a hyperscaler’s pricing whims or capacity shortages. This Capex focus essentially allows a business to build its own “private engine room,” which is the only way some enterprises can maintain their desired pace in this high-velocity environment.

Organizations are currently managing dozens of AI pilots simultaneously using a mix of closed, open, and SaaS models. How do you prevent technical debt when moving these proofs of concept into full-scale production?

With nearly 50% of enterprises currently juggling more than 30 AI pilots, the risk of creating a fragmented, unmanageable “spaghetti” of systems is incredibly high. To prevent this, we must implement a unified orchestration layer that acts as a translator between closed models and open-source derivative tools. I recall a case where a firm was using a closed model for customer-facing chat and a derivative open model for internal documentation, only to find they couldn’t share data between them without a total rebuild. The key is to standardize the data “plumbing” before you ever launch the pilot, ensuring that the model—whether it’s a packaged SaaS or a custom build—is interchangeable. If you don’t enforce these standards early, the cost of “unifying” these 30+ pilots in 2028 will likely eclipse the cost of the infrastructure itself.

Soaring token volumes are a primary driver of the need for increased compute capacity and specialized architecture. What practical steps can technical teams take to optimize these volumes to keep costs manageable?

Token volumes are doubling and tripling at a rate that can quickly bankrupt an unmonitored project, so the first step is implementing “token quotas” and aggressive prompt optimization. Technical teams should look toward “small language models” for specific, repetitive tasks, which can handle workloads at a fraction of the compute cost of a massive general-purpose model. This trend is a massive tailwind for the “neocloud” market, which is projected to reach $400 billion in revenues by 2031 because these providers offer specialized, high-density compute that hyperscalers often can’t match for specific AI tasks. Choosing a neocloud provider over a traditional hyperscaler is often a decision based on the need for raw, specialized power—like the gigawatts of TPU capacity we see companies like Anthropic securing. It’s about moving from a “general store” cloud to a “precision factory” infrastructure.

The scale of AI spending is forcing a blur between business strategy and technical infrastructure decisions. How should leadership roles evolve to ensure that massive capital expenditures align with long-term value?

The era of the CIO working in a vacuum is over; today, every major infrastructure purchase is a core business strategy decision that involves the CEO and CFO from day one. Leadership must evolve so that the CFO understands the nuances of GPU depreciation while the CIO understands the long-term market positioning of the company’s data assets. The consequences for failing to integrate these roles are dire: you end up with “stranded assets” or hardware that doesn’t solve a business problem, leading to massive write-downs. We are seeing a new type of executive leadership where technical literacy is no longer optional, but the very foundation of financial planning. When budgets are tripling, a technical mistake isn’t just a glitch—it’s a threat to the organization’s solvency.

What is your forecast for AI infrastructure?

My forecast is that by 2028, the “buy vs. rent” debate will be settled in favor of a “sovereign hybrid” model where 70% of large enterprises own their core AI compute engines. We will see the $400 billion neocloud market become the primary battleground for specialized talent and hardware, leaving traditional public clouds to handle generic web services. As token volumes continue to explode, the companies that thrive will be those that treated their infrastructure as a proprietary asset rather than a utility bill. Ultimately, the successful enterprise of 2030 will look less like a software company and more like a high-tech utility, managing its own vast reservoirs of compute and data with the same precision as a power grid. Infrastructure will finally be recognized as the ultimate competitive moat.

Trending

Subscribe to Newsletter

Stay informed about the latest news, developments, and solutions in data security and management.

Invalid Email Address
Invalid Email Address

We'll Be Sending You Our Best Soon

You’re all set to receive our content directly in your inbox.

Something went wrong, please try again later

Subscribe to Newsletter

Stay informed about the latest news, developments, and solutions in data security and management.

Invalid Email Address
Invalid Email Address

We'll Be Sending You Our Best Soon

You’re all set to receive our content directly in your inbox.

Something went wrong, please try again later