Why NVIDIA is 3 years ahead of everyone - NVIDIA CEO explains | Jensen Huang and Lex Fridman
Jensen Huang explains how NVIDIA stays 3 years ahead by anticipating AI innovation through internal research, industry collaboration, and maintaining flexible architecture like CUDA. He describes how they predicted the shift from LLM inference to agentic systems that use tools, leading to the design evolution from Grace Blackwell to Vera Rubin racks.
Summary
Jensen Huang outlines NVIDIA's strategy for staying ahead in AI hardware development, emphasizing the challenge of hardware cycles taking 3 years while AI model architectures change every 6 months. NVIDIA addresses this through three key approaches: conducting internal basic and applied research to gain hands-on experience, working with every AI company globally to understand industry challenges and trends, and maintaining CUDA's flexible architecture that balances specialization with adaptability. Huang provides concrete examples of successful prediction, such as anticipating mixture of experts architectures leading to MVLink 72 development, and the evolution from Grace Blackwell racks designed for LLM inference to Vera Rubin systems designed for agentic computing. He explains that predicting these shifts comes down to logical reasoning about what AI systems need to become useful digital workers - they must access files, conduct research, and use existing tools rather than reinventing everything. Using the analogy of a humanoid robot that would use a microwave rather than beam microwaves from its fingers, Huang argues that AI agents will naturally integrate with existing computing infrastructure, leading to what he calls a reinvention of the computer itself.
Key Insights
- NVIDIA maintains competitive advantage through three strategies: internal research for hands-on experience, collaboration with every AI company globally to understand industry challenges, and maintaining CUDA's flexible architecture that balances specialization with adaptability
- Hardware architectures evolve every 3 years while AI model architectures change every 6 months, requiring companies to anticipate what will happen 2-3 years in the future
- NVIDIA anticipated the shift from LLM inference to agentic systems, evolving from Grace Blackwell racks focused on processing LLMs to Vera Rubin racks with storage accelerators and new components designed for agents that use tools
- AI agents will function as digital workers that need to access ground truth through file systems, conduct research, and use existing tools rather than having everything built-in, because waiting for universally smart AI before making it useful is impractical
- The development of agentic AI systems that use existing tools and infrastructure represents a reinvention of the computer, with profound implications for the future of computing
Topics
Full transcript available for MurmurCast members
Sign Up to Access