China Tech Talk

Episode 5: AI infrastructure


Listen Later

This week, I invited my friend Simon on the show to talk about AI infrastructure. This is a very interesting topic to think about. We talked about the reason for Nvidia’s market dominance and the problems that Intel and AMD have faced in developing their own market place. We talked about the number of chips involved and the challenges in building large data centers. Here is a chart of the Hopper deliveries for 2024. Keep in mind that Google also has their own TPU for inference.

Here is a chart of Chinese Hyperscaler’s purchases of H20 in 2024. But keep in mind that they do rent compute from public/state built data centers that use smuggled in H100/H200 as well as Ascend chips.

Here is the deliveries of H100/H800 in 2023. So if we just consider the Hopper deliveries from 2023 and 2024, the American hyperscalers have quite the computation advantage over Chinese hyperscalers. Although, the gap is much smaller if we factor in all the smuggled in chips. As I’ve discussed before, China had them everywhere to the point where they were sitting idle in many cases.

I also asked Simon about the amount of chips needed to do inference. How was Tencent able to have enough compute for 8 million DeepSeek R1 requests at the same time with much less compute than what Google had. He sent me this chart of where the results improve logarithmically in compute for reasoning. So if you look at green line, going from 4 generations to 16 generations improved accuracy by about 15% and then going from 16 generations to 64 generations improved the result by another 5%. If we look at Google search results, they appear to be giving progressively better AI results on top. That is likely from running through and generation more tokens on their reasoning models.

So you can choose to serve 8 million prompts at same time with vastly less compute, but you will also have inferior results. although you will also reach diminishing returns pretty soon.

Here is a chart of rental cost from various Nvidia chips in China. Despite the increased demand for inference post DeepSeek, the cost of all Nvidia rental cost continue to drop, so they are likely not facing a crunch for computation yet.



This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit tphuang.substack.com
...more
View all episodesView all episodes
Download on the App Store

China Tech TalkBy TP Huang

  • 5
  • 5
  • 5
  • 5
  • 5

5

4 ratings


More shows like China Tech Talk

View all
The Tim Ferriss Show by Tim Ferriss: Bestselling Author, Human Guinea Pig

The Tim Ferriss Show

16,183 Listeners

Odd Lots by Bloomberg

Odd Lots

1,957 Listeners

Sinica Podcast by Kaiser Kuo

Sinica Podcast

611 Listeners

The Duran Podcast by The Duran

The Duran Podcast

608 Listeners

ManifoldOne by Steve Hsu

ManifoldOne

97 Listeners

Ones and Tooze by Foreign  Policy

Ones and Tooze

352 Listeners

Multipolarity by Multipolarity

Multipolarity

40 Listeners

Geopolitical Cousins by Jacob Shapiro & Marko Papic

Geopolitical Cousins

76 Listeners