My personal take-away from the call is that Nvidia is far more than just an AI silicon company. There was a lot of talk about various services being provided, mostly on the cloud. I’m still digesting the announcements and descriptions of:
Sovereign AI Clouds:
CEO Huang said “You’re seeing sovereign AI infrastructure. People now recognize that they have to utilize their own data, keep their own data, keep their own culture, process that data, and develop their own AI. … The number of sovereign AI clouds that are being built is really quite significant. And my guess is that almost every major region, and surely every major country, will have their own AI clouds.”
Nvidia AI Foundry for the “development and tuning of custom generative AI enterprise applications running on Azure. Customers can bring their domain knowledge and proprietary data and we help them build their AI models using our AI expertise and software stock in our DGX cloud, all with enterprise grade security and support. SAP and Amdocs are the first customers of the NVIDIA AI foundry service on Microsoft Azure.”
DGX cloud service
“Our monetization model is that with each one of our partners they rent a sandbox on DGX Cloud, where we work together, they bring their data, they bring their domain expertise, we bring our researchers and engineers, we help them build their custom AI. We help them make that custom AI incredible. Then that custom AI becomes theirs. And they deploy it on the runtime that is enterprise grade, enterprise optimized or outperformance optimized, runs across everything NVIDIA. We have a giant installed base in the cloud, on-prem, anywhere.”
Nvidia AI Enterprise Factory customers pay monthly for access to a sandbox and services from the NVIDIA team to help them build and train custom generative AI models. The enterprises can then deploy on NVIDIA AI Enterprise for $4500 per GPU per year.
Back to hardware, Nvidia is doing more than just GPUs, like AI-specific tuned networking (InfiniBand). And even on the compute side, it’s worth understanding a bit about the various “SuperChips” that Nvidia sells. Huang pointed out that one of their SuperChips, the HGX “Hopper” H100 has 35,000 parts and weighs 70 pounds. It’s more of a board than a chip, no wonder each costs tens of thousands of dollars.
While Nvidia is expecting to have its latest AI data center GPU, the Hopper H200 ready to ship by Q2 next year (and it’s 2X faster than the current H100), Nvidia also has a “Grace” data center chip, based on ARM. They’ve got a board/SuperChip that has two dies containing 144 ARM cores. The latest Apple M2 processor, for comparison, has just 8 ARM cores. It’s interesting to me that while ARM’s initial success was initially tied to lower power usage than Intel’s x86, which of course is important for battery powered phones and tablets, now Nvidia is advertising low power consumption for data center CPUs. Data centers are getting so large that saving 2X on power consumption and infrastructure costs, with the same or better performance is apparently a big deal.
As you may recall, Nvidia wanted to buy ARM the company, but the UK nixed that deal. Still, Nvidia is obviously a big licensee and that they’re coming out with an data center oriented ARM CPU shows Nvidia isn’t even just an AI play. And it introduces new data center architectures to enterprises. Nvidia has a “Grace Hopper” combination SuperChip that has a Grace ARM-based CPU paired with one or more Hopper GPUs for the data center. Even heavy AI workflows have non-AI compute requirements, so this combination should prove compelling.
Nvidia’s AI software, CUDA, remains a huge moat as it enables customers to upgrade Nvidia hardware with little to no changes in their software, and means switching to any other product, even if they wanted to, would be a big deal.
A profitable high-growth megacap that even pays a dividend. Imagine that.