I guess the history of AI and Nvidia’s role in its development aren’t well known, even by those in the industry. Nvidia’s AI dominance isn’t some sudden flash in the pan.
Arguably, the Deep Learning Revolution began 14 years ago, when Professor Geoffrey Hinton and some grad students won the 2012 ImageNet competition with AlexNet, achieving an image-recognition accuracy that had never been seen before, using a new kind of Neural Net technology, built on Nvidia hardware.
According to Hinton, AlexNet would not have happened without Nvidia. Thanks to their parallel processing capabilities supported by thousands of computing cores, Nvidia’s GPUs — which were created in 1999 for ultrafast 3D graphics in PC video games, but had begun to be optimized for general computing operations — turned out to be perfect for running deep learning algorithms.
“In 2009, I remember giving a talk at NIPS [now NeurIPS] where I told about 1,000 researchers they should all buy GPUs because GPUs are going to be the future of machine learning,” Hinton told VentureBeat last fall.
(From How Nvidia dominated AI — and plans to keep it that way as generative AI explodes | VentureBeat )
It was clear parallel processing was the path forward for AI. So, why didn’t Intel and AMD invest in GPUs and AI a decade ago? After all, Hinton told everyone a decade and a half ago that GPUs were the key. Was Huang a genius, or was everyone else not paying attention, or just not interested?
Nvidia was pushing non-graphics compute on GPUs well before 2012, releasing the first version of CUDA in 2007, which is what enabled researchers to program the GPUs for AI computing. This is Nvidia realizing that their GPUs could do more than just video games, and enabling that to happen. In 2007, with development on that started sooner, of course.
.
Nvidia is exactly a company that came up with something brand new that is taking the competition years to catch up. Meanwhile, Nvidia keeps pumping out innovations and new products.
Intel remains in world of hurt, doing both design and fabrication. AMD caught Intel flat-footed in the x86 world, but that’s so last couple of decades. ARM eat Intel’s lunch in the mobile space, and AMD doesn’t play in the ARM stadium.
This isn’t just who has the fastest chip. Nvidia mostly sells CPUs on boards or in modules (up to 8 per “tray”) that use NVLink for fast interconnection between them. From the article above:
“While other players offer chips and/or systems, Nvidia has built a strong ecosystem that includes the chips, associated hardware and a full stable of software and development systems that are optimized for their chips and systems,” analyst Jack Gold wrote for VentureBeat last September.
But, even on the basis of chips, Intel and AMD are just now shipping units they claim are as fast as Nvidia’s Hopper (which Nvidia disputes). In a few months, Nvidia will be shipping Blackwell, which is 4X faster (30X in the ARM-pairing “Grace Blackwell” configuration, while using ¼ the power). And when will the next generation of AMD and Intel AI chips come to market? It’s been a two year cycle at best for them, while Nvidia is now on a one year cycle.
One thing that’s different today than the x86 cycles of last century is that speed matters. People were willing to save money on home or even business PCs that had AMD processors even if they were slower. But, with AI the time to market isn’t just computer scientists writing code and compiling, then testing, there’s a whole cycle of training. How long does it take to feed the entire internet into your NN (or transformer, to keep things current), then test the results? Tesla has been spending Billions of dollars on Nvidia chips, and has just recently not been “compute bound” as Musk called it, training on who knows how many miles of car camera data.
Jensen Huang on the Nvidia ER call:
“Let me give you an example of time being really valuable, why this idea of standing up a data center instantaneously is so valuable and getting this thing called time to train is so valuable. The reason for that is because the next company who reaches the next major plateau gets to announce a groundbreaking AI. And the second one after that gets to announce something that’s 0.3% better. And so the question is, do you want to be repeatedly the company delivering groundbreaking AI or the company delivering 0.3% better?”
@ThisShallPass is correct that competition is going to try really hard to to do the “your margin is my opportunity” thing that happens when one company dominates so lucratively. The problem appears to be that this is currently a space where even the fastest isn’t fast enough, and getting your completed trained model out there first is important in terms of recognition and therefore revenue. So, the days when AMD could make a good business with an x86 chip that was 85% as fast as Intel but 25% of the cost, don’t appear to be mapping to today’s AI world. At least not yet.