Americas

  • United States

Nvidia chips sold out? Cut back on AI plans, or look elsewhere

News
Nov 20, 20258 mins

There may be better options for enterprises struggling to access enough Nvidia chips to deliver on their AI strategy.

Nvidia high-performance chip technology
Credit: gguy / Shutterstock

Nvidia CFO Colette Kress’s claim that “The clouds are sold out, and our GPU-installed base […] is fully utilized,” may have thrilled shareholders listening to the company’s earnings call on Wednesday, but it’s bad news for CIOs and data center managers who were counting on Nvidia for increased AI computing capacity, as they will have to change suppliers — or change their plans.


CEO Jensen Huang, asked during the same earnings call whether he saw a realistic path for supply to catch up with demand over the next 12-18 months, claimed that all was going to plan, saying, “We’ve done a really good job planning our supply chain. Nvidia’s supply chain basically includes every technology company in the world.”

Not everyone is convinced. Among the doubters is Forrester senior analyst Alvin Nguyen, who said Thursday he has had clients ask what to do when addressing the lack of available Nvidia GPUs, since demand far exceeds supply.

There are, he said, “other options for both on-premises (AMD, Intel, custom ASICs, CPUs) and in the cloud (TPUs, custom ASICs), although demand means these options may not be enough for everyone’s AI ambitions.”

Nguyen added, “for enterprises and their CIOs, not being able to get the AI infrastructure needed to achieve your full AI vision means re-evaluating those ambitions and paring them back to what is possible. Leveraging AI platforms and services others (Salesforce, ServiceNow, and so forth) provide can help mitigate some needs.”

‘Everything is moving so fast.’

Senior IT executives, he said, should also consider looking at working with smaller models with reduced infrastructure needs, and experimenting with them to help inform future AI decisions.

The constant innovations in this space, said Nguyen, “may help enterprises build or mitigate technical debt depending on what and when they decide on their infrastructure. I know this is an ‘it depends’ answer, but everything is moving so fast that the answers are only clear in hindsight.”

Matt Kimball, principal analyst at Moor Insights & Strategy, said the question about Nvidia and the availability of GPUs is a good one. “Some of the challenges I see organizations face are quite avoidable if some thought exercise around right-sizing of infrastructure were to take place,” he noted.

He pointed out, “Nvidia chips (or any chips for that matter) have different performance profiles, as well as different performance per watt and performance per dollar profiles. The latest [Nvidia] GB300 is not always the right fit for the job. And when we split between training and inference, this approach to rightsizing the solution for the need is even more critical.”

By doing this, said Kimball, “organizations will find they are less reliant on a latest generation chip that has a long line of (much larger) customers queued up and waiting on delivery. The other exercise is to consider whether the Nvidia chip is always necessary for your needs. I know this may sound like tech heresy. Still, especially as it pertains to inference, it is beneficial to understand what the inference environment looks like, where the infrastructure is being deployed, and what the workload entails.”

It may very well be a case where an ASIC-based solution is a better fit for that real-time sensor-driven environment on an oil rig, for example, he observed. 

Kimball added, “I’m certainly not saying not to acquire Nvidia. Still, regardless of supply chain issues, it is a very good exercise to think about your AI needs holistically and align the right-sized acceleration to your needs.” 

In addition to that, he said, “the cloud is always a choice. These are the first-in-line customers for Nvidia silicon, and it is perfectly natural to leverage the cloud for AI needs.” 

The key for CIOs: Be proactive

Gaurav Gupta, VP analyst at Gartner, had this advice: “While Nvidia continues to say that they have a strong control of their supply chain, the complexity is so high that it should be a priority for CIOs to monitor.”

He added, “not only are there potential shortages on some of the big known aspects, like leading-edge wafers, advanced packaging, and HBM [high bandwidth memory], but in my opinion it is the unrecognized constraints for smaller components and precision machinery parts for thermal management, liquid cooling, and server racks that could be bottlenecks. Plus, everyone needs to plan for power availability to run these data centers.”

The key, said Gupta, is to be proactive, plan ahead, and “not be the last in the queue” when it comes to ordering compute resources such as GPUs.

Scott Bickley, advisory fellow at Info-Tech Research Group, noted, “the world is starting to question how Nvidia is going to move from an [about] $250 billion annualized run rate to $350 billion, and eventually to $500 billon plus. Yet they clearly stated every available chip is sold out, in fact, that they could have sold more if they had the inventory. It is reasonable to apply scrutiny to this incredibly complex and fragile supply chain and ask the question, ‘What happens if there is a material failure that limits GPU shipments?’”

Supply chain risks, he said, “are numerous in nature; however, it is clear that Nvidia is customer Number One with all of their suppliers, which drives an inordinate allocation of resources to ensure that production flows. Any disruption would likely be materials-based as opposed to a process or labor issue from their vendor base.”

He added, “geopolitical events would be the most likely origin of any type of medium to long term disruption, think China-Taiwan, expansion of the Russia-Ukraine conflict, or escalation in the US-China trade war.”

For lower impact events, he said, “[Nvidia] does a nice job of setting conservative shipment goals and targets for Wall Street, which they almost invariably beat quarter after quarter. This provides some cushion for them to absorb a labor, process, or geopolitical hiccup and still meet their stated goals. Shipment volumes may not exceed targets, but shipments would continue to flow; the spice must flow after all.”

In a worst-case scenario where shipments are materially impacted, there is little recourse for enterprises that are not large-scale cloud consumers with clout with the limited providers in the space, Bickley added.

Enterprises joining a ‘very long queue’

According to Sanchit Vir Gogia, the chief analyst at Greyhound Research, the Nvidia earnings call “confirms that the bottleneck in enterprise AI is no longer imagination or budget. It is capacity. Nvidia reported $57 billion in quarterly revenue, with more than $51 billion from data center customers alone, yet still described itself as supply-constrained at record levels.”

Blackwell and Blackwell Ultra, he said, have become the default currency of AI infrastructure, yet even at a build rate of roughly 1,000 GPU racks per week, the company cannot meet demand.

Long-term supply and capacity commitments, said Gogia, “now stand at around $50.3 billion, and multi-year cloud service agreements have jumped to $26 billion, implying that much of the next wave of capacity has already been pre-booked by hyperscalers and frontier labs. Enterprises are not stepping into an open market. They are joining the back of a very long queue.”

The supply imbalance, he said, “is not just about chips. It is about everything that has to wrap around them. The filings point to long manufacturing lead times, tight availability of advanced packaging and high bandwidth memory, and significant prepayments and non-cancellable commitments to secure future capacity.”

Gogia also suggested that the single biggest decision for CIOs now “is whether to design their AI strategy around Nvidia or around the risk of Nvidia. These are not equivalent positions. To design around Nvidia is to accept that the platform is the gold standard and lean into it by placing orders 12 months ahead, using multiple OEMs for the same configuration, coordinating with finance teams on prepayments, and building programme timelines that can absorb shipment shifts.”

To design around the risk, he said, “is to recognize that Nvidia is essential but cannot be the only path, and to treat diversification as a resilience measure rather than a philosophical debate.”

Paul Barker is a freelance journalist whose work has appeared in a number of technology magazines and online, including IT World Canada, Channel Daily News, and Financial Post. He covers topics ranging from cybersecurity issues and the evolving world of edge computing to information management and artificial intelligence advances.

Paul was the founding editor of Dot Commerce Magazine, and held editorial leadership positions at Computing Canada and ComputerData Magazine. He earned a B.A. in Journalism from Ryerson University.

More from this author