I wanted to expand on something I mentioned in passing on an earlier post that makes me very excited about NVDA – their inference microservices, called NIMs. I’m posting this because I haven’t seen a lot of mentions on the board of this, but I find it quite relevant to mention.
NIMs are software AI snippets that a developer can use to “drop” AI into their code. They are like old-timey software library modules (such as “print” etc.) but for AI. In NVDA’s words: “It expands the developer pool by abstracting away the complexities of AI model development…”
What excites me about this is if NVDA gets this right, these NIMs will take off. They will take off because why would any software developer reinvent the wheel when they can focus on the rest of the car?
If a developer uses the NIMs, this ties the solution/product to the NIMs, bringing an additional revenue source for NVDA – possibly massive and ongoing - that is software-based.
In addition, NVDA is trying to create an AI “marketplace” where companies/developers can add their own NIM-like services for sale (likely new development on top of existing NIMs) and offer them too. Remains to be seen if it actually takes off, as NIMs are very new, just announced in March.
NIMs are free to use for development and research but cost once you deploy: $4500 per GPU per year, or $1 per GPU per hour.
So here’s an example – remember NIMs were just announced in March (although I imagine some companies had early access):
- NVIDIA Morpheus … enables developers to create optimized applications for filtering, processing, and classifying large volumes of streaming cybersecurity data. Morpheus incorporates AI to reduce the time and cost associated with identifying, capturing, and acting on threats, bringing a new level of security to the data center, cloud, and edge.
And this mention on Seeking Alpha about a Zscaler and NVDA partnership:
(For those who don’t subscribe, here is one relevant paragraph:
Zscaler (ZS) will also leverage Nvidia Morpheus framework and Nvidia NIM inference microservices to deliver additional predictive and generative AI solutions to market, including the Zscaler ZDX Copilot with NVIDIA Morpheus and the Zscaler ZDX Copilot with Nvidia NIM.)
So this enables ZScaler to deliver additional AI capabilities in Cybersecurity quickly. And it locks ZS onto NVDA GPUs and additionally paying NVDA for the use of these software AI capabilities.
There’s also NIMs for healthcare (Clara), audio and video (Maxine), recommenders (Merlin), LLM-related stuff like text to speech and speech recognition (Riva) and I’m sure more to come.
If NVDA puts as much energy into this as they put into their chips and they get it right, this software will be instrumental to AI applications running everywhere. Equally important, it’s solidifying the moat and adding an additional annual revenue stream from software.