Skip to content
WWT LogoWWT Logo Text (Dark)WWT Logo Text (Light)
The ATC
Ctrl K
Ctrl K
Log in
What we do
Our capabilities
AI & DataAutomationCloudConsulting & EngineeringData CenterDigitalImplementation ServicesIT Spend OptimizationLab HostingMobilityNetworkingSecurityStrategic ResourcingSupply Chain & Integration
Industries
EnergyFinancial ServicesGlobal Service ProviderHealthcareLife SciencesManufacturingPublic SectorRetailUtilities
Learn from us
Hands on
AI Proving GroundCyber RangeLabs & Learning
Insights
ArticlesBlogCase StudiesPodcastsResearchWWT Presents
Come together
CommunitiesEvents
Who we are
Our organization
About UsOur LeadershipLocationsSustainabilityNewsroom
Join the team
All CareersCareers in AmericaAsia Pacific CareersEMEA CareersInternship Program
Our partners
Strategic partners
CiscoDell TechnologiesHewlett Packard EnterpriseNetAppF5IntelNVIDIAMicrosoftPalo Alto NetworksAWSGoogle CloudVMware
What we do
Our capabilities
AI & DataAutomationCloudConsulting & EngineeringData CenterDigitalImplementation ServicesIT Spend OptimizationLab HostingMobilityNetworkingSecurityStrategic ResourcingSupply Chain & Integration
Industries
EnergyFinancial ServicesGlobal Service ProviderHealthcareLife SciencesManufacturingPublic SectorRetailUtilities
Learn from us
Hands on
AI Proving GroundCyber RangeLabs & Learning
Insights
ArticlesBlogCase StudiesPodcastsResearchWWT Presents
Come together
CommunitiesEvents
Who we are
Our organization
About UsOur LeadershipLocationsSustainabilityNewsroom
Join the team
All CareersCareers in AmericaAsia Pacific CareersEMEA CareersInternship Program
Our partners
Strategic partners
CiscoDell TechnologiesHewlett Packard EnterpriseNetAppF5IntelNVIDIAMicrosoftPalo Alto NetworksAWSGoogle CloudVMware
The ATC
Spend OptimizationResearchResearch NoteHigh Performance Computing (HPC)ATCSupply Chain and Integration ServicesInfrastructure ServicesConsulting ServicesWhat we do
WWT Research • Research Note
• March 19, 2026 • 9 minute read

Beyond the Shortage: A CPO's Playbook for Navigating the AI Infrastructure Crunch

AI infrastructure crunch is structural, not cyclical, driven by memory, storage and power constraints. CPOs must act now — by right‑sizing technology, extending asset lives, restructuring contracts and operationalizing supplier risk — to protect margins and secure capacity through at least 2028.

In this report

  1. Why this time really is different
  2. Three forces you need to understand
    1. 1. The HBM wafer problem
    2. 2. The storage crisis is worse than it looks
    3. 3. The power wall nobody wants to talk about
  3. Six moves to make before this gets worse
    1. 1. Right-size technology to the actual workload.
    2. 2. Consolidate storage density and use data reduction aggressively.
    3. 3. Stretch asset life cycles to seven years.
    4. 4. Restructure your commercial agreements.
    5. 5. Make supplier risk management operational, not ceremonial.
    6. 6. Treat software efficiency as a procurement lever.
  4. Looking past the crisis
  5. Sources

March 2026 |  World Wide Technology |  Infrastructure Advisory Practice                    

The signs have been visible for two years. The semiconductor supply chain was drifting toward a structural break, and that break has now arrived. What the market is experiencing is not a cyclical dip that corrects in a quarter or two. It is a fundamental reallocation of global manufacturing capacity toward AI, and it will define infrastructure economics through at least 2028.

CPOs who act on this in the next 12 months will protect their margins and lock in capacity while it's still available. The ones waiting for the market to "normalize"? They'll be bidding against each other for scraps at peak pricing.

What follows is an analysis of the forces driving the crunch, why it differs from previous cycles, and six specific moves procurement leaders should consider making now.

Why this time really is different

There's a statistic that should keep every infrastructure procurement leader up at night: Compute performance has outpaced memory bandwidth by roughly 500% over the past decade. That gap is the root cause of nearly everything happening in the semiconductor market right now.

And here is the part that deserves more attention than it gets. The AI performance gains celebrated by the industry? They came almost entirely from software innovation, not from better chips. NVIDIA's own analysis breaks down a thousandfold GPU performance improvement over ten years. The biggest contributors were smarter number representation (16x) and complex instructions (12.5x). Moore's Law, the force that historically rescued procurement budgets every refresh cycle? It delivered 2.5x. That's it.

The implication is blunt: the industry cannot build its way out of this shortage with incremental fab capacity. The bottleneck is at the physics level. Chip fabrication improvements are decelerating. Power consumption is heading in the wrong direction. And memory technology, which is the actual enabler of AI workloads, has been chronically underinvested for years because DRAM vendors were stuck in commodity pricing cycles that crushed their R&D budgets.

The numbers on the ground confirm this. Q4 2025 semiconductor inventory indices dropped 5.1% quarter-over-quarter, with memory and foundry segments now in severe shortage territory. Server memory and storage configurations have already seen price surges of 150% to 300%. PC prices are projected to climb up to 20% in the first half of this year. And in a twist that perfectly illustrates how broken the market is, legacy DDR4 memory is now more expensive than DDR5 in several regions because vendors are announcing end-of-life and everyone is scrambling.

Three forces you need to understand

1. The HBM wafer problem

High-bandwidth memory (HBM) is the silicon that makes AI accelerators work. The next generation, HBM4, is a genuine architectural leap: double the bus width, triple the bandwidth and 16-high die stacking. But the production economics are punishing. HBM consumes more than 2.5 times the wafer capacity of standard DRAM because the dies are over 50% larger and yields are significantly lower. By 2027, HBM is expected to eat roughly a quarter of all DRAM wafer capacity worldwide.

What this means in practice is that every HBM chip that gets manufactured is a DDR5 chip that doesn't. Your organization may not be building AI training clusters, but you are absolutely competing for the same underlying wafer supply as every hyperscaler on the planet. There is no "non-AI" lane in this supply chain. There's just one lane, and it's congested.

Figure 1: HBM's growing share of DRAM wafer starts and revenue, 2022–2028

2. The storage crisis is worse than it looks

NAND flash vendors cut fab utilization by about 10% in early 2025 to push prices up after a stretch of weak profitability. Then AI demand exploded. Meanwhile, the "nearline" HDD market, which handles a lot of AI inference storage, has effectively seized up. Lead times for high-capacity drives hit 52 weeks. That's a year. The unmet HDD demand is now spilling into the NAND/SSD market, creating a secondary wave that flash suppliers simply can't absorb. Enterprise SSD prices are expected to rise 70% to over 100% this year, and the undersupply extends well into 2027.

3. The power wall nobody wants to talk about

AI chips are on track to exceed 1 kilowatt of thermal design power by the end of this year. To put that in perspective, a conventional server CPU runs at 40 to 130 watts. That is a 10x increase in heat generation per chip, multiplied across thousands of GPUs per data center, with corresponding demands for liquid cooling infrastructure, structural building reinforcement and electrical capacity that most facilities were never designed for.

A server you can buy but can't power or cool generates zero value. Power and cooling capacity now belong on the critical path for infrastructure procurement, right next to component availability.

Six moves to make before this gets worse

The strategic landscape is daunting, but the tactical playbook is actionable. The framework below maps the three market forces to six specific procurement responses.

             Figure 2: Strategic Immunization Framework — six moves to insulate procurement from AI-driven volatility

1. Right-size technology to the actual workload.

This sounds obvious and almost nobody does it well. Stop deploying PCIe Gen5 NVMe drives for Tier II business applications when Gen4 or QLC technology will do the job at a fraction of the cost. Look at your DIMM configurations: are you running 32x32GB DIMMs when 16x64GB DIMMs would deliver equivalent performance at lower total cost? At current pricing, these are not marginal savings. They're material. Reserve premium silicon for the workloads that actually require it, whether that's AI training, high-frequency trading, or whatever genuinely demands top-tier performance.

2. Consolidate storage density and use data reduction aggressively.

Moving from standard 4TB drives to high-capacity enterprise QLC SSDs in the 30TB to 120TB range means fewer drives, fewer controllers and lower power consumption. Combine that with storage arrays capable of 4:1 data reduction through compression and deduplication, and you effectively quadruple usable capacity without buying more hardware. One important caveat: Bigger drives mean a bigger blast radius when one fails. Make sure your controllers can handle rapid rebuilds.

3. Stretch asset life cycles to seven years.

The three-to-five year server refresh cycle is a relic of predictable pricing. It no longer makes sense. Extending to seven years is designed to bridge the gap to 2028, when industry forecasts project prices will fall roughly 50% as supply catches up. Use third-party maintenance providers to keep assets running past OEM support windows. This is not austerity. It is disciplined capital deployment at a moment when the market is punishing impatient buyers.

4. Restructure your commercial agreements.

Move to cost-plus pricing that locks in discount percentages off list prices. Push for "bill and hold" agreements that secure today's rates for inventory you'll need later. Track open market component pricing independently, because server vendors have been known to charge nearly double what the open market commands when buyers lack visibility. Diversify your memory supplier base now, not after your primary vendor misses a delivery.

Consumption-based models are also worth a hard look. Storage-as-a-service or cloud options can shift spend from capital to operating budgets, which avoids capitalizing assets purchased at peak pricing. It's not right for every workload, but the optionality has real value when you can't predict whether a component will cost 2x or 3x what it did last year.

5. Make supplier risk management operational, not ceremonial.

About half of supply chain organizations report that their business continuity management was effective during their most recent disruption. Half. That's a coin flip on whether your plans actually work when you need them. On the other hand, organizations that invest in proactive risk mitigation consistently report measurably better profit outcomes.

The gap between "having a plan" and "having a plan that works" comes down to a few specific things: 

  • Segment suppliers by value at risk, not just spend. That means the revenue exposed if they fail and how hard they are to replace.
  • Classify risks into categories that drive different responses: some warrant terminating the relationship outright, some need active monitoring, and some require hands-on mitigation with business leadership involved.
  • Run tabletop exercises with your critical vendors so you actually test whether their recovery teams can execute.
  • Map your critical supply paths at least two tiers deep, because a disruption at a packaging substrate manufacturer you've never heard of can shut you down just as fast as a problem at your direct supplier.
Figure 3: Value at Risk supplier segmentation model — prioritize by business impact, not spend

6. Treat software efficiency as a procurement lever.

This one gets overlooked in every procurement strategy we review, and it shouldn't. When every additional gigabyte of memory costs 2x to 3x what it did 18 months ago, and every additional watt of power requires cooling infrastructure you may not have, software efficiency translates directly to dollars. Organizations that optimize their AI inference pipelines for memory utilization, rationalize their storage hierarchies and clean up inefficient code will simply need less hardware. In this market, needing less is the most dependable form of supply assurance.

Looking past the crisis

It's tempting to treat this purely as a cost problem. Prices are up, budgets are strained and boards are asking uncomfortable questions. But the CPOs who will come out ahead are framing it differently. They see a window to build procurement capabilities and supplier relationships that will outlast the shortage. Technology tiering discipline, real supplier risk management, software-driven efficiency — none of these expire when prices come back down. These are permanent upgrades to how your procurement function operates.

The market will find a new equilibrium, probably around 2028. HBM yields will improve. NAND investment will start catching up. New fab capacity will come online. But between now and then, the gap between organizations that moved early and those that waited will only widen.

The window for action is right now. Not next quarter. Now.

Sources

WWT Internal & External Research 

Assess Supplier Business Continuity Management for Risk Mitigation 

Proactive Supplier Risk Management: A Strategic Approach 

Semiconductor 2Q25 Update: Trends & Impact on Supply Chains 

Improve Supplier Risk Measures to Capture Cost-Effectiveness 

Rising Costs Ahead: Managing AI-Driven Price Increases in Data Center Infrastructure 

Semiconductor Inventory Analysis Worldwide, 4Q25 

Emerging Tech: AI Vendor Race: Semiconductor Technology Limits GenAI Scaling 

WWT Research
Insights powered by the ATC

This report may not be copied, reproduced, distributed, republished, downloaded, displayed, posted or transmitted in any form or by any means, including, but not limited to, electronic, mechanical, photocopying, recording, or otherwise, without the prior express written permission of WWT Research.


This report is compiled from surveys WWT Research conducts with clients and internal experts; conversations and engagements with current and prospective clients, partners and original equipment manufacturers (OEMs); and knowledge acquired through lab work in the Advanced Technology Center and real-world client project experience. WWT provides this report "AS-IS" and disclaims all warranties as to the accuracy, completeness or adequacy of the information.

Contributors

Wesley Palmer
Managing Director
Ryan Wickham
VP & General Manager, Solutions Infrastructure
Rob Prokop
Principal Solution Architect

Contributors

Wesley Palmer
Managing Director
Ryan Wickham
VP & General Manager, Solutions Infrastructure
Rob Prokop
Principal Solution Architect

In this report

  1. Why this time really is different
  2. Three forces you need to understand
    1. 1. The HBM wafer problem
    2. 2. The storage crisis is worse than it looks
    3. 3. The power wall nobody wants to talk about
  3. Six moves to make before this gets worse
    1. 1. Right-size technology to the actual workload.
    2. 2. Consolidate storage density and use data reduction aggressively.
    3. 3. Stretch asset life cycles to seven years.
    4. 4. Restructure your commercial agreements.
    5. 5. Make supplier risk management operational, not ceremonial.
    6. 6. Treat software efficiency as a procurement lever.
  4. Looking past the crisis
  5. Sources
  • About
  • Careers
  • Locations
  • Help Center
  • Sustainability
  • Blog
  • News
  • Press Kit
  • Contact Us
© 2026 World Wide Technology. All Rights Reserved
  • Privacy Policy
  • Acceptable Use Policy
  • Information Security
  • Supplier Management
  • Quality
  • Accessibility
  • Cookies