The AI boom has a clear headline: spend massively on chips, data centres and power. Yet beneath that gleaming façade is a growing inefficiency — a huge share of the most expensive AI accelerators are simply waiting for work. The result is a costly mismatch: premium hardware that can compute at blistering speed but sits idle because the networks, storage and orchestration that feed it haven’t kept pace. That imbalance is rapidly shaping where companies invest next: not only in GPUs, but critically in the pipes and systems that keep them busy.
Why the chips are idle
High‑end GPUs used for training and inference can cost $20,000–$30,000 apiece and consume roughly a kilowatt of power when active. But Jürgen Hatheier of Ciena lays out a stark metric: in many hyperscale environments, GPUs spend roughly half their time idle, waiting for data or coordination signals to arrive. Multiply that idleness across tens of thousands of accelerators in a single cluster and the waste becomes enormous — both financially and energetically.
The network is now the critical path
For years the industry accepted that compute improvements would outpace networking for a while. That window, however, has narrowed to a dangerous gap. The innovation curve for accelerators has moved much faster than that for high‑speed connectivity and the storage systems needed to keep them fed. The consequence is simple: owning the best chip is insufficient; what matters is how fast you can supply it with data and how effectively machines can collaborate.
Enter a wave of investment in networking and edge infrastructure. Fibre routes, low‑latency interconnects, pre‑deployed capacity and better data‑locality strategies are now strategic differentiators. Operators are pre‑laying links and cloud providers are redesigning campus layouts and inter‑datacentre fabrics to remove the stalls that idle expensive compute. Investors are noticing — network equipment vendors have seen significant stock gains as markets price this shift.
Storage, CPUs and the broader supply chain matter too
It’s not just the network. Storage systems capable of streaming data at multi‑terabyte per second rates, and CPU fleets able to orchestrate millions of tasks, are essential. Many AI workloads depend on fast parameter servers, sharded model states and synchronous updates that put pressure on all layers of the stack. Where any single layer lags, the accelerators sit idle.
Enterprises and sovereign demands complicate matters
Hyperscalers (Meta, Microsoft, Alphabet) have led the spending spree, but enterprises are increasingly establishing private GPU clusters for sovereignty, latency and data‑control reasons. These on‑prem environments intensify the problem because many organisations don’t have the networking or operational expertise to maintain high utilisation at scale. Therefore, inflating numbers of expensive racks may further stress capacity unless firms simultaneously invest in connectivity and orchestration.
What success looks like
High utilisation of AI hardware requires a holistic approach:
Providers that combine these elements will extract the most value from their GPU investments. The marginal returns of a new accelerator are not just measured in teraflops per second, but in how quickly it can be kept busy.
Where investment will go next
Expect capital expenditure to shift balance. While GPUs and specialised accelerators remain headline items, the next wave of investment will favour high‑capacity fibre, advanced switching fabrics, flash and storage tiers tuned for AI, and sophisticated CPU and orchestration stacks. Companies that underestimate the need for end‑to‑end optimisation risk paying for idle hardware and losing the economic advantage of AI at scale.
In short, compute is no longer the sole star of the AI story. The real winner will be the organizations that recognise AI as a systems challenge — one where network, storage, CPU orchestration and facility readiness matter as much as the chips themselves.
