A few days ago, Jensen Huang was a guest on the BG2 podcast. Iâve heard him speak many times before, but this conversation struck me as extraordinary. We get an almost unfiltered view of Nvidia, but even more so of the trajectory of AIâs development as a technology and an industry.
On my way to Dubai, I had a chance to listen to the whole thing. This is one of the most interesting conversations Iâve listened to in a while and here Iâm going to share my favourite bits1 and takeaways⌠Enjoy!
The flywheelâs flywheel
Jensen spoke about how the stack of technologies that comprise AI is accelerating. Nvidia is focused on the rate of change of that acceleration⌠at its heart, the flywheel of machine learning helps them achieve 2-3x more performance every year.
Many people used to believe that designing a better chip with more FLOPs, more bits and bytes, was the key. Youâd see their keynote slides full of charts and specs. And yes, horsepower does matter. But thatâs old thinking.
In the past, software was staticâjust an application running on Windows. The way to improve it was to make faster chips. But machine learning isnât human programming; itâs not just about the software. Itâs about the entire data pipeline. The real key is the machine learning flywheel.
The most important part is enabling data scientists and researchers to be productive in this flywheel. It starts at the very beginning. People often donât realise that it takes AI to curate data to teach an AI, and that process is incredibly complex. With smarter AI curating the data, we now even have synthetic data generation, adding more ways to prepare data for training. Before you even get to training, you have massive amounts of data processing involved. [âŚ]
Every step along the flywheel is challenging. In the past, we focused on making things like Excel or Doom faster. But now, you need to think about how to make the entire flywheel faster. [âŚ]
In the end, the exponential rise comes from accelerating the whole system.
Accelerating the entire system requires a holistic approach that considers Amdahlâs law. Amdahlâs law states that the overall system speedup is limited by the fraction of the system that cannot be parallelised or improved. To achieve significant acceleration, you need to optimise every component of the AI pipeline, from data preparation to inference, not just focusing on individual steps like training.
Amdahlâs law suggests that if a component in a process takes 30% of the total process time and I accelerate that component by 3x, I didnât really accelerate that process by that much.
(âThat muchâ being only 20% improvement in the overall system speed.)
[âŚ] you really want to create a system that accelerates every single step because only in doing the whole thing can you really materially improve that cycle time and that flywheel. The rate of learning in the end is what causes the exponential rise.
This is Nvidiaâs mission. To do all this requires an integrated ecosystem.
Nvidiaâs ecosystem approach
Huang describes how the supply chain is working together to increase performance on a per dollar and per watt basis at one or two orders of magnitude faster than the equivalent time of Mooreâs law.
The important thing is that Mooreâs law was a social contract. It was an agreement across the range of suppliers in the semiconductor industry to step up to meet Intelâs public roadmaps. Nvidia is doing something similar but at a much larger scale.
Keep reading with a 7-day free trial
Subscribe to Exponential View to keep reading this post and get 7 days of free access to the full post archives.