For years, I was a believer that Autonomy needed billions of miles of data for AI training. I no longer believe that to be the case.
Here’s a video from a channel on movie effects:
If you don’t know about Green Screens, you should watch it from the beginning, but essentially what Hollywood often does is shoot actors in front of a green screen, then use computers to replace green with a computer-generated background. The problem is at the edges of the subjects, especially with things like hair and motion blur, individual pixels will not be entirely green nor entirely subject, and the computer “replace color” algorithm doesn’t know what to do with somewhat green pixels - are they part of the subject (tree green, not screen green), or part-subject and part-background (which would mean partial transparency)?
What they decided to do (and the head of Avatar’s special effects company didn’t think it would work), was to use a commonly available rendering engine to render realistic scenes against a green screen, but since they know what is subject and what is background screen, they were able to train an AI accordingly.
Well, not that many 18-hour training sessions later (and remember this channel doesn’t have millions of dollars or access to thousands of Nvidia chips), they were successful. It’s open source, so hopefully more people in the industry will pick this up and run with it.
My point in posting here is that they didn’t need billions of hours of actual footage shot in front of a green screen, they didn’t even need millions of hours - and I infer that they didn’t even have tens of thousands of hours. Because they knew where the problem areas were (hair, swords, glasses of water, veils, etc.) they were able to concentrate their synthetic data creation accordingly and be efficient in training.
I think this is probably what Nvidia is doing for their AV stack, going into Mercedes, Nuro, and even Lucid in the future. They don’t need 10 Billion miles like Elon claimed, they just need the right set of, say, ten-thousand synthetically-generated miles, for which Nvidia can generate and train in probably 10 minutes given their resources.
Tesla pioneered E2E autonomy back in late 2022 (according to Walter Isaacson’s book). Mobileye’s CEO wrote a blog post saying it was garbage. Waymo adopted some E2E and so did Nvidia (with less fanfare). And now we see Nvidia’s demo during CES just a couple months ago doing really well, especially considering how late they started.
I do believe Tesla is getting there, and is probably ahead in terms of driving capabilities and ability to expand, but I don’t believe Waymo will fall further behind and I believe Nvidia might be catching up. Waymo and Nvidia have the problem that they’re not automotive manufacturers and so physically scaling will be harder for them, but Waymo has Hyundai building cars, and OEMs can adopt Nvidia’s system relatively easily (and cheaply, it’s open source) now.
Nvidia may be making things more difficult for themselves by adding LiDAR, but I suspect OEM pressure requires them to do so even though it’s not necessary and does little more than slow the creation of synthetic data down while increasing training effort and time.
But, my main point here is that the only issue with synthetic data (at least for cameras) is knowing what to simulate. And while there are literally an infinite number of situations that can crop up, I think that coming up with a complete-enough set of situations to render for training is easier/faster than just brute-force capturing 10 Billion miles.
In short, I believe Elon’s wrong, but that doesn’t mean Tesla won’t have other advantages with autonomy.
On a separate note, there was a open NHTSA meeting yesterday:
https://www.nhtsa.gov/events/av-public-meeting-2026
Zoox and Waymo presented, I think Tesla was in attendance.
They discussed changes in FMVSS for AVs. News article here:
It’s now open for public comment. The goal is to let these companies have vehicles on public roads without manual controls. Zoox already has a wavier for around 2500 vehicles (something NHTSA has done in the past for developmental vehicles), but all 3 companies want to put many more vehicles than that on the roads.
I know someone who attended the meeting and spoke with a Tesla attendee, who told him that the Cybercab should directly conform to the updated FMVSS.


