Huang Isn't Planning to Wake Up a Loser
You may have already listened to Jensen Huang on the recent episode of Dwarkesh podcast. If you haven’t, I highly recommend you listen to the entire episode and not just the bits and pieces clips. This wasn’t the kind of episode that you want an AI to summarize either.
What made the episode so enjoyable to listen to is I thought Dwarkesh did an excellent job in eliciting Huang’s opinions on a range of topics that are hotly debated. Frankly speaking, I came to the conversation with a slight bias to Huang’s point of view of selling chips to China, but came away thinking Huang’s arguments weren’t convincing enough.
Here on MBI Deep Dives, I am more interested in micro economics concerns such as competitive dynamics, moats etc as I consider the broader macro, geopolitical concerns are largely beyond my paygrade even though they can be quite impactful at times. There were plenty of materials in the episode that were very much in my narrow micro economics interest even though most people are mostly talking about the debate around selling chips to China.
Speaking of competitive dynamics, the below excerpt on Huang’s opinions about Trainium and TPUs are particularly worth highlighting:
I would welcome Trainium to demonstrate their 40% that they claim all the time. I would love to hear them demonstrate the cost advantage of TPUs. It makes no sense in my mind. It makes absolutely zero sense. On first principles, it makes no sense.
So I think the reason why we’re so successful is simply because our TCO is so great. Secondly, you say 60% of our customers are the top five, but most of that business is external. For example, most of Nvidia in AWS is for external customers, not internal use. Most of our customers at Azure, obviously all of our customers are external. All of our customers at OCI are external, not internal use. The reason why they favor us is because our reach is so great. We can bring them all of the great customers in the world. They’re all built on Nvidia. And the reason why all these companies are built on Nvidia is because our reach and our versatility is so great.
So I think the flywheel is really install base, the programmability of our architecture, the richness of our ecosystem, and the fact that there’s so many AI companies in the world. There’s tens of thousands of them now. If you were one of those AI startups, what architecture would you choose? You would choose an architecture that’s most abundant. We’re the most abundant in the world. You’d choose the one that has the largest installed base. We’re the largest install base. And you’d choose the one that has a rich ecosystem.
So that’s the flywheel. That’s the reason why, between the combination of: one, our perf per dollar is so great that they have the lowest cost tokens. Second, our perf per watt is the highest in the world. So if one of these companies, if our partners, built a one gigawatt data center, that one gigawatt data center better deliver the maximum amount of revenues and number of tokens, which directly translates to revenues. You want it to generate as many tokens as possible, maximize the revenues for that data center. We are the highest tokens per watt architecture in the world. Lastly, if your goal is to rent the infrastructure, we have the most customers in the world. So that’s the reason why the flywheel works.
Jensen correctly notes that Amazon (AWS), Microsoft (Azure), and Google (GCP) buy mountains of GPUs primarily because their external cloud tenants demand them. Hyperscalers, simplistically speaking, are increasingly acting as capital-intensive distributors for Nvidia. It may take a while, but I still wonder unless we get past this compute constrained environment, the true competitive dynamics between ASICs and Nvidia GPUs may be hard to ascertain. In the current environment, you will probably sign a contract whoever can offer you compute. As a result, Trainium, TPUs, and GPUs all can coexist and thrive in this environment. I am noticing a growing consensus among investors that “we may never have enough compute” which implies worrying about competitive dynamics in a relatively compute abundance scenario largely irrelevant. I, however, lean towards believing in the power of capitalism and there is hardly any shortage that capitalism has failed to eliminate over time. As a result, while it’s excruciatingly challenging to pinpoint a timeline when we will reach such relative compute abundance scenario, I suspect we will eventually have to deal with such question. If you value these businesses assuming such question will never arrive but it does arrive five years from now, that has much more valuation implications than the most investors may imagine.
I continue to think that in a relative compute abundance scenario, TPU is the most protected precisely because Google has enormous captive internal workloads: Search query serving, YouTube ranking, ad targeting, Gemini across every surface, Waymo, and Workspace AI features. Even if external TPU demand goes to zero, internal workload volume alone justifies continued investment and tape-outs. They’re obviously not immune, just less exposed than Trainium and Nvidia’s GPUs. My guess is DeepMind would be very happy to get higher compute allocation if Google Cloud faces any difficulty for selling compute to external customers at attractive ROIC.
Huang also made an interesting point about ASICs vs GPU margins:
“don’t forget, even in ASICs margins are really quite high. Nvidia’s margin is 70%, let’s say. But ASIC margins are 65%. What are you really saving?
You’ve got to pay somebody. I think the ASIC margins are incredibly good, from what I can tell. They believe it too. They’re quite proud of their incredible ASIC margins.”
I found this argument to be a bit misleading. Epoch AI suggests that Blackwell chips sell for $30k to $40k whereas the bill of materials (BOM) costs only ~$5.7k to $7.3K which implies an eye-watering 82% margin. They, however, did point out that “Since most Blackwell revenue comes from servers and rack-scale systems, which may carry lower margins, NVIDIA’s realized margins on Blackwell sales may be lower than these chip-level estimates.”

When a hyperscaler pays Nvidia $30k+ for a Blackwell unit, Nvidia's margin is captured on the entire bill of materials plus the software stack. Broadcom's or Marvell's "65% ASIC margin" is on a much narrower slice. The hyperscaler pays TSMC directly for wafers and pays SK Hynix, Samsung, or Micron directly for HBM, which together are the majority of the BOM on a current-generation accelerator. Broadcom's margin sits on top of its contribution, not on top of the whole chip. Saying "65% vs 70%" makes it sound like near-parity, but the 65% is applied to perhaps a quarter of the system cost while the 70% is applied to all of it.
A hypothetical chip with ~$6k of real BOM cost sold by Nvidia at $30k captures $24k of margin per unit. The same chip produced through a design-services relationship with Broadcom might carry ~$8k of Broadcom margin plus the same ~$6k BOM, landing the hyperscaler at $14-15k all-in. Even if Broadcom's percentage margin on its slice is as high as Nvidia's, the dollar savings per unit are enormous because Broadcom's slice is so much smaller.
Custom silicon programs do, however, carry real costs that doesn’t get surfaced in this simple framing. There is internal engineering headcount, inference and training framework support, and the opportunity cost of engineers who could be working on revenue-generating products instead. These are fixed and amortized, but they are not zero, and for a program that fails to reach sufficient volume, the amortized per-unit cost can be meaningful vs just merely comparing the BOM. This is the actual reason most companies do not build custom silicon because you need Google-scale or AWS-scale internal demand AND a top-tier engineering team to make the amortization math work. For the handful of operators where the math does work, the savings are not 65% vs 70%. They are likely closer to single-digit-thousands vs tens-of-thousands per chip, and I’m sure Huang knows this.
However, there are different layers to this debate. Gavin Baker pointed out that while model portability used to make investors think model developers might gain leverage over the chip suppliers over time, that may be changing with Blackwell and even more so with Rubin. From Baker’s post (emphasis mine):
As system level architectures diverge (torus vs. switched scale-up topologies, memory hierarchies, networking primitives), true portability is eroding. The Mi300 and Mi325 had roughly the same scale-up domain size as Hopper while Blackwell’s scale-up domain is 9x larger than the Mi355 scale-up domain, etc.
Many frontier models are now being explicitly co-designed for inference on specific hardware like GB300 racks. Codex on Cerebras is another example. Those models run less efficiently on other systems and the performance differentials will only widen. A model that runs well on Google’s torus topology will run less efficiently on Nvidia’s switched scale-up topology and vice versa - the data traffic is fundamentally different as a byproduct of the models being parallelized across the different topologies.
Google’s internal teams - and increasingly the Anthropic teams as they become the most important customer of almost every cloud - have the luxury of operating across the stack (models, chips, networking) - but that is not the case for the rest of the market and other prospective users. Anthropic is the exception, not the rule. To wit, Anthropic and Google allegedly have a mutual understanding where Anthropic can hire the TPU engineers they need every year to ensure that they can continue to get the most out of the TPU.
Given the overwhelming importance of cost per token to the economics of the labs, models will be run where they run best. Most extremely large MoE models will run best on GB300s given the importance of having a switched scale-up network like NVLink for MoE inference. When training was the dominant cost for labs and power was broadly available, labs were optimizing to minimize capex dollars. Model portability was a way to create leverage over suppliers. I think that drove a lot of the focus on portability.
Today, inference costs as measured by tokens per watt per dollar are everything. Inference is way more important than training costs (inference is effectively now part of training via RL). Labs are therefore now optimizing for inference. This means increasing co-design and higher go-forward switching costs for individual models between systems. I do think this explains why Anthropic and Nvidia came together: Anthropic needed Blackwells and Rubins to inference at least *some* of their models economically. And Mythos might just end up being released coincident with the availability of Rubins for inference.
TLDR: as labs shift their focus from training to inference, the costs of portability and the upside of co-design to maximize tokens per watt per dollar both rise. Portability is likely to begin decreasing as a result.
Reading Baker’s post made me think that the hyperscalers will have to pay the “Nvidia tax” whether they like it or not; as long as the hyperscaler customers see material benefit for using Nvidia GPUs over hyperscaler ASICs, hyperscalers hands will be tied. And if hyperscalers appear unwilling or hesitant to pay such exorbitant taxes, Nvidia will be happy to allocate more of their chips towards neoclouds, and many customers, especially AI startups will likely follow. Always remember, “You’re not talking to somebody who woke up a loser.”
In addition to “Daily Dose” (yes, DAILY) like this, MBI Deep Dives publishes one Deep Dive on a publicly listed company every month. You can find all the 67 Deep Dives here.
Current Portfolio:
Please note that these are NOT my recommendation to buy/sell these securities, but just disclosure from my end so that you can assess potential biases that I may have because of my own personal portfolio holdings. Always consider my write-up my personal investing journal and never forget my objectives, risk tolerance, and constraints may have no resemblance to yours.
My current portfolio is disclosed below: