>
>
Big Tech’s $700B AI spending surge hides the real problem: turning chips into working capacity

News
Big Tech’s $700B AI spending surge hides the real problem: turning chips into working capacity
AI demand isn’t slowing down in 2026. If anything, it’s outpacing what the biggest cloud companies can actually deliver, and that gap is driving a massive wave of spending.
Here’s what’s going on. Customers want more AI computing power than the cloud can reliably provide today. So companies are pouring billions into infrastructure. Not because demand is weak, but because meeting that demand is harder than it looks.
Buying chips is the easy part. Turning those chips into a working AI service is where things get complicated. A chip only makes money once it’s fully deployed inside a data center with enough power, cooling, networking, and software to keep everything running smoothly.
That’s why this has turned into a race. It’s no longer just about who can buy the most GPUs. It’s about who can turn all that spending into usable AI capacity the fastest.
This also explains why investors feel uneasy. Companies have to spend upfront because building out infrastructure takes time. But investors want to see results now. They’re looking for clear signs like faster cloud growth, rising AI revenue, or stronger ad performance.
So the story in 2026 comes down to one thing. Big cloud companies are still trying to close the gap between how much AI customers want and how much AI they can actually deliver. Until that gap shrinks, the spending isn’t going anywhere.
Why spending is still rising before the money fully shows up
The spending only looks excessive if you assume the shortage is almost over. The numbers suggest it is not. Recent reports about massive AI investment plans show that Alphabet, Amazon, Meta, and Microsoft are continuing to scale aggressively, reinforcing the idea that capacity is still far behind demand.
Bridgewater estimates Alphabet, Amazon, Meta, and Microsoft will spend about $650 billion on AI-related infrastructure in 2026, up from about $410 billion in 2025. Alphabet has said its 2026 capital spending could reach $175 billion to $185 billion. Amazon has projected $200 billion, up from $131 billion in 2025. Meta expects $115 billion to $135 billion. Microsoft spent $37.5 billion in capital expenditures in the December 2025 quarter alone. Those are not normal expansion numbers. They suggest these companies still believe they are short of the AI capacity their customers want.
The companies’ own comments point the same way. Alphabet said Google Cloud remained in a tight supply environment and expected capacity constraints through 2026. Amazon said AWS demand remained strong despite capacity constraints. Microsoft said Azure growth would have been above 40% if it had used newly available GPUs differently. Put plainly, the problem is not finding customers. The problem is getting enough working AI capacity online fast enough to serve them.
That is why the spending is showing up before the full payoff does. The companies are not only buying chips. They are also paying for data centers, power equipment, cooling systems, network gear, and all the other pieces needed to make those chips useful. In effect, they are buying future supply before all the revenue has arrived.
Investors understand that up to a point. But they are no longer willing to cheer huge AI spending just because it sounds ambitious. Reuters showed that shift in January: investors were more comfortable with big AI spending when a company could show that it was already helping growth or had a believable path to stronger profits. Meta got more room because AI was clearly helping its ad business. Microsoft faced more pressure because its cloud growth did not rise enough to make the spending feel clearly worthwhile.
The problem is bigger than just getting enough chips
For a while, the AI story was mostly about chip shortages. That is still part of the picture, but it is no longer the whole story. Even as companies like Alibaba push to develop new chips to meet rising demand, the bottleneck has shifted toward integrating those chips into fully functioning systems.
Now the problem looks more like a chain. First a company has to get the chips. Then it has to have a data center ready. Then it needs enough power. Then cooling. Then networking. Then the software systems that decide how the machines are used. Only after all of that can the company turn the hardware into a reliable product.
That is why buying more GPUs and creating more AI capacity are no longer the same thing. A company can secure more chips on paper and still take months to make them truly useful. In 2026, the real test is how fast spending turns into working AI service customers can actually use.
A simple way to judge that is to ask five questions. How quickly can a company bring new capacity online? How limited is it by power? Does it have the networking parts it needs? Can it direct valuable AI workloads onto the new systems quickly? And can it keep those systems busy enough to earn a solid return? Those questions say more than chip totals do, because they show whether spending is turning into business.
Power is the clearest example. On March 19, 2026, Reuters reported that Google had signed utility contracts that would make up to 1 gigawatt of its data-center electricity demand available for cutbacks during times of peak power use. That is a sign that electricity is no longer just a background cost. It is now part of the AI race itself. A company does not make deals at that scale unless power has become a serious operating issue.
Networking shows the same thing in a different way. On March 24, 2026, Broadcom said supply problems were reaching beyond semiconductors into lasers and printed circuit boards used in optical transceivers, with lead times for some PCB parts stretching from about six weeks to six months. Chips are still tight. But Broadcom’s warning matters because it shows that even if a company gets enough accelerators, it can still be slowed by the parts that connect those systems together. This is reinforced by Nvidia’s ongoing adjustments to global chip supply and approvals, which highlight how availability alone does not guarantee immediate usability.
The longer-term energy picture makes this even harder to ignore. EPRI said in February 2026 that data centers could consume 9% to 17% of U.S. electricity generation by 2030. The IEA’s base case projects global data-center electricity use at roughly 945 terawatt-hours by 2030, slightly above Japan’s current electricity consumption. Those numbers matter because they suggest power and efficiency are becoming lasting competitive issues, not temporary side problems.
Which companies look best positioned?
These companies are not all dealing with the same problem in the same way.
Amazon’s case is the easiest to understand. Its projected $200 billion in 2026 capital spending looks like a response to AWS demand that still exceeds the capacity it can provide. That makes Amazon look less like a company chasing hype and more like one trying to meet customer demand before rivals do. Its challenge is not proving there is a market. It is building enough working AI capacity without hurting returns too much. This growing AI push is also influencing Amazon’s broader strategy, including renewed hardware efforts tied to AI capabilities.
Alphabet is one of the clearest examples of the argument in this article. It said Google Cloud would remain constrained through 2026, guided to $175 billion to $185 billion in capex, said about 40% of technical infrastructure capex goes to data centers and networking, and signed utility-flexibility agreements at very large scale. That suggests Alphabet is not just buying compute. It is trying to secure all the supporting pieces needed to make that compute usable.
Meta shows that even giant spending does not solve everything. It plans $115 billion to $135 billion in capital spending, but also said those costs include payments to third-party cloud providers and acknowledged internal capacity constraints through much of 2026. That is important because it shows that when speed matters, even a company spending heavily may still need to rent outside capacity.
Microsoft shows the investor side most clearly. It suggested Azure growth would have been above 40% with different GPU allocation. That supports the idea that demand is still strong and supply is still limiting results. But it also shows why investors are more demanding now. If spending keeps climbing and growth still does not show enough of a lift, markets start to treat that spending less like future strength and more like current risk.
What really matters now
The market is no longer rewarding spending by itself. It is rewarding how quickly spending turns into working service.
A company that secures more chips but then waits on substations, cooling systems, or network parts has not solved the real problem yet. A rival that gets fewer chips but brings them online faster and fills them with valuable AI workloads may get better results. In this stage of the AI buildout, speed matters. Revenue follows working capacity, not procurement headlines.
That is especially important because building useful AI systems is not just about infrastructure. It is also about whether the software and workflows actually work in real-world conditions. Many organizations are discovering that AI systems can struggle outside controlled environments, reinforcing the idea that usable capacity is more than just hardware availability.
That is where the risk sits too. Chips still matter. Advanced foundry capacity is still tight. AWS and Microsoft are still talking about constraints. But once the chips are bought, the fight becomes operational. Can the company power the systems, cool them, connect them, manage them, and keep them productive? If not, the spending starts to look less like a moat and more like pressure on margins.
Alphabet has already warned that rising infrastructure investment will materially increase depreciation and data-center operating costs such as energy. Meta’s operating margin has already come under pressure as capex has outpaced revenue growth. Those are important warning signs. They show what happens when the cost of building shows up before the business payoff does.
Big cloud companies are spending so much on AI in 2026 because customers still want more AI service than the clouds can reliably deliver. The scarce thing is not just chips. It is working capacity. The companies that win will be the ones that turn spending into reliable, usable AI service fastest.
P.S. Here are the 3 biggest takeaways from Nvidia 2026 GTC.
Recommended Articles



