You are currently viewing SemiWiki as a guest which gives you limited access to the site. To view blog comments and experience other SemiWiki features you must be a registered member. Registration is fast, simple, and absolutely free so please, join our community today!
Open Source bi-weekly convo w/ Bill Gurley and Brad Gerstner on all things tech, markets, investing & capitalism. This week, Brad and Clark Tang sit down with Jensen Huang, founder & CEO of NVIDIA, for a sweeping deep dive on the new era of AI. From the $100B partnership with OpenAI to the rise of AI factories, sovereign AI, and protecting the American Dream—this episode explores how accelerated computing is reshaping the global economy. NVIDIA, OpenAI, hyperscalers, and global infrastructure: the AI race is on. Don’t miss this must-listen BG2.
A few standouts for me:
* Moving general purpose computation to accelerators vs CPUs - SQL/Snowflake/Databricks and associated processing on NVIDIA ?
* Yearly new chip/hardware generations is a huge differentiator. They use AI to speed.
* Extreme co-design - chip/software/rack/system/datacenter all developed concurrently
* ASIC vs CPU vs GPU -
- Rubin CPX (long context processing, diffusion video generation accelerator) is precursor for other application specific specialized accelerators.
- Maybe a data processing app specific chip/subsystem next
- Transformer architecture still changing rapidly - programmability still required.
- Only real system-level AI chip competition is Google/TPU
- ASICs only useful for mid-volume - too much gross margin given up for middleman. Smart NICs, and Transcoders are good candidates for ASICs. Not a good option for fundamental compute engine for AI, where underlying algorithms are changing regularly.
- Data centers / AI factories are a soup of ASICs and other chips - need to be orchestrated and co-developed with supply chain.
- NVIDIA targeting lowest Total Cost of Ownership at data center level. Someone could offer ASIC chips at zero $$ and still be less economical. Tokens per gig and tokens per watt are compelling.
* NVLink Fusion and Dynamo leading the way in creating next-gen open AI solutions and associated ecosystem.
* Not just a chip company. The AI infrastructure company.