For this week’s episode, I spoke with Chris Miller, the author of Chip War, about the rise of Nvidia.
While OpenAI gets the lion’s share of the public adulation for the sudden excitement about generative intelligence, Nvidia’s H100 chips are powering much of the generative AI frenzy. Nvidia’s stock has climbed over 200% over the past 12 months. And the company has become a key investor in generative AI startups.
Miller (who comes on the show around the 41-minute mark) talks through Nvidia’s history and the geopolitical war raging over the production of chips.
In the first part of the episode, Cerebral Valley AI Summit co-hosts Max Child, James Wilsterman, and I discuss how big technology companies are working to fend off this new generation of AI startups.
Give it a listen
Get full access to Newcomer at www.newcomer.co/subscribe
00:00:09
Welcome back to Cerebral Valley. I'm here with Max Child and
00:00:13
James Woolsterman, my friends and Co founders of Volley.
00:00:17
This episode we are talking about chips and big tech.
00:00:21
My interview, after a conversation with Max and James
00:00:25
is with Chris Miller, the author of Chip War, The fight for the
00:00:30
World's most critical technology.
00:00:32
He is an expert. We talk a lot about NVIDIA.
00:00:36
Yeah, so stick around for that. Max James welcome back.
00:00:40
Nice to be here. Eric, Glad to be here.
00:00:42
The starting point question that I wanted to frame things up with
00:00:45
is, do you think GPU capacity, the quality of these graphics
00:00:49
processing units and what NVIDIA is putting out is the main
00:00:53
driver for the sort of generative AI revolution we're
00:00:57
seeing right now? What do you mostly attribute it
00:01:00
to? The research papers, The new
00:01:02
approaches. How?
00:01:04
How do you? How do you?
00:01:05
Who do you give? Credit.
00:01:07
I guess I give more credit to the research papers.
00:01:11
Not that I don't discount what NVIDIA has done here.
00:01:16
I just, I think they were in the right place at the right time
00:01:19
with the right technology to accelerate what was possible
00:01:23
through due to the research. But yeah, the research was the
00:01:27
breakthrough. I do think there was some
00:01:29
intentionality. I mean, it's always hard to tell
00:01:31
with AI when there's so much. It's a hypey thing to talk
00:01:35
about. And this is ACEO that also made
00:01:37
a bunch of money off crypto mining.
00:01:39
So happy to dive into wild speculation.
00:01:42
Clearly I don't know. Max, you have a take here.
00:01:45
I mean I think I agree with James.
00:01:46
I just think that, you know even four or five years ago we were
00:01:50
already starting to see the beginnings of this with GPT 2
00:01:52
and you know earlier versions of the the foundational models,
00:01:55
right. Also in areas outside of large
00:01:58
language models you saw big progress in, you know our hobby
00:02:01
horses of speech synthesis, speech recognition and other
00:02:04
areas. So I think that even with five
00:02:06
year old GP US, you still could pretty to do pretty cool stuff
00:02:10
with the new research that was outright whereas.
00:02:12
If the research never comes out, I don't think that you're making
00:02:15
this kind of progress in, in any of these kind of large large
00:02:18
language model areas that we or or any of these areas of the
00:02:21
stack that we've seen so far. So right.
00:02:24
I mean, chips, even though there's potentially exponential
00:02:28
improvement, it's still like, you know, on a trajectory,
00:02:32
whereas a paper sort of comes out of nowhere, gives people a
00:02:35
new approach and then sort of revolutionizes what's possible.
00:02:40
I mean, I think a different way to think about it would be like,
00:02:42
OK, GPT has gotten like 1000 times better in the last three
00:02:46
years and like the chips haven't gotten 1000 times better in the
00:02:49
last three years. So like, what's driving that?
00:02:51
Like, it's probably the software and then the people like
00:02:54
figuring out how to make GPT, right?
00:02:55
So yeah, I just think it's clearly based on the ideas in
00:02:59
the software. The chips are great, no doubt,
00:03:01
but it's not the key driver. I I do think it so NVIDIA has
00:03:06
clearly gotten rich off this stuff.
00:03:08
It is now a $1.1 trillion company.
00:03:11
In one year the stock is up 258% and in five years it's up 564%.
00:03:21
So this, this has been a great run for the company.
00:03:27
I mean what do you, what do you guys make of NVIDIA investing in
00:03:32
all these companies? I mean similarly if it's into,
00:03:37
you know Amazon, Google, like all these companies are making
00:03:41
investments. The money that they're spending
00:03:43
obviously comes back to their business, right.
00:03:45
NVIDIA wants AI to be super active, so puts out money that
00:03:50
gets spent on NVIDIA chips. Yeah.
00:03:52
What do you make of, I mean it's round, it's round tripping in
00:03:56
the OR it's you know spending money to make money, but it's it
00:04:00
doesn't seem as sort of negative as in some cases.
00:04:04
I mean, isn't the investment NVIDIA is putting out into the
00:04:07
ecosystem like a rounding error compared to their revenue or
00:04:09
compared to their profits? I mean, like, I I mean I think
00:04:12
that they're making such ludicrous quantities of money
00:04:14
that kicking back like 1 to 2% into the startup community that
00:04:18
might invest in NVIDIA chips in the future is, is obviously just
00:04:21
a very good business practice, right.
00:04:23
Right. Well, it seems like they're
00:04:24
literally playing kingmaker, right.
00:04:26
I mean like companies like Core Weave and who are sort of chip
00:04:30
doing chip adjacent stuff depend on their access to H1 Hundreds
00:04:35
and a lot of these foundation models seem like they're
00:04:37
fundraising on the premise that hey we have access like this is
00:04:43
differentiated partially because we have the deal to get GPU
00:04:47
access when other people don't. And so there's this hope that
00:04:51
you know, you're you have a Moat in that you have access to to
00:04:54
chips that other people don't. The scarcity today of the best
00:04:58
in class chip, the H 100 is. Has to be a temporary situation.
00:05:02
I just like the economics of like making more of these chips
00:05:05
is a very good idea for NVIDIA. So they're going to do it.
00:05:07
But they they clearly want to build like a Amazon Web Services
00:05:12
type competitor. And so having eyes and ears in a
00:05:15
bunch of startups and seeing maybe who they could acquire to
00:05:18
do that is smart. I do think the interesting
00:05:20
question again is like Intel basically had like 2 decades of
00:05:24
like total monopolistic dominance of of PC chips, right?
00:05:29
And. The reason basically was that
00:05:31
it's like really, really, really hard to build fabrication plants
00:05:35
to to build CP US. And it took a long time for like
00:05:39
TSMC and other people to sort of catch up on that front and then
00:05:42
eventually kind of surpass them in many ways.
00:05:43
But the question I guess for GP US I think is like, is it that
00:05:47
technically hard to make GP US that like whatever the next
00:05:51
version of the H 100 is? You know that they're going to
00:05:54
be on to that before anyone's even caught up to the H 100, and
00:05:57
so they'll have this insurmountable 2 two to
00:05:59
three-year advantage where if you want the best AIGPUS, it's
00:06:02
always going to be NVIDIA for the next 20 years.
00:06:04
Because that seems like a plausible case to me.
00:06:06
Do you think Microsoft was smart to invest 10 billion in open AI
00:06:09
given it's clearly going to distract the company as has been
00:06:13
reported? I think my friend in the Wall
00:06:15
Street Journal reported as much that they directed sort of their
00:06:18
attention to open AI away from Microsoft.
00:06:21
And and part of what you're saying is they they are giving
00:06:26
up access to their own GPUs to Open AI.
00:06:29
Is that accurate to to say so? Yeah, I mean, I think it was.
00:06:34
I think it was a good decision and.
00:06:36
I agree. Definitely a good decision.
00:06:39
I I can't imagine they would be, as you know critical to where we
00:06:45
are in the in the AI era right now had they not made that
00:06:48
decision they. Have you know early access to
00:06:53
all the technology that Open AI is creating.
00:06:56
They got GPT 4 into Bing before it was in public in you know in
00:07:02
front of customers for in front of Open AIS customers and seems
00:07:06
to be continuing with what they are trying to do with Office and
00:07:11
Windows, you know the Office suite.
00:07:14
So I think. Just that early access to Open
00:07:18
AI products seems worth the investment.
00:07:21
I think, I think the sort of meta question you're asking that
00:07:24
I think is really interesting is like.
00:07:26
Is big tech going to win the AI race or whatever, right?
00:07:29
Like is like, I mean we always used to talk about Fang right?
00:07:32
Like Facebook, Amazon, Apple used to be Netflix and Google
00:07:34
right? You could replace the N and Fang
00:07:36
with NVIDIA and it would be perfect.
00:07:37
So Fang OK and like, goodbye streaming.
00:07:40
Hello. Chips.
00:07:41
Yeah, Goodbye streaming. Hello.
00:07:42
Chips. Right.
00:07:43
And I love Netflix, So they never got over like a couple 100
00:07:45
billion, which is where you said, you know, NVIDIA is a
00:07:47
trillion now. So pathetic. 200 Three $100
00:07:51
billion company. Yeah.
00:07:53
Is fan going to win AI? Like, I think that's a really
00:07:56
interesting question. And I think like I'm kind of
00:07:59
leaning, yes, I guess to your point, like it seems like
00:08:02
Microsoft's been able to grab, you know, Open AI, which is, per
00:08:06
our last podcast, still with the most valuable asset in the
00:08:08
entire industry, right. It seems like Amazon's getting
00:08:12
very cozy with Anthropic, in case that's actually the other
00:08:15
big asset in the foundation model category.
00:08:17
Facebook is absolutely like ripping on open source models.
00:08:21
You know, Llama. They're going to put like LLMS
00:08:24
inside their ad products so you can have automated ads on
00:08:27
Facebook and Instagram and all that stuff, you know?
00:08:29
They're also putting, they're also putting LLMS into consumer
00:08:35
chat. You know, products, right?
00:08:36
You can chat with celebrities, and I'm sure they're doing a lot
00:08:39
with image generation. Yeah, exactly.
00:08:42
And then Google seems to be like fully ripping on, just like
00:08:44
dumping large language models into every single Google
00:08:46
product. Like Google.
00:08:47
You know, Google Docs, Google Photos, I don't know if that is
00:08:51
the conventional wisdom. And so on.
00:08:53
OK. Google ripping for Google.
00:08:55
For the record, you have to like grade them on the Google curve,
00:08:58
which is they're in a competent company coming out with new
00:09:01
ideas. Given that Google has not come
00:09:03
out with a new idea in 15 years, they are doing a tremendous job
00:09:06
putting AI products into their existing, into their existing
00:09:10
products, right? And so, yeah, for them, the
00:09:13
number of launches this year has been tremendous, right, Graded
00:09:16
on the Google curve, right. And then Apple has basically
00:09:18
done nothing. But Apple always does things
00:09:21
like 2 years late and then hopefully they get it right.
00:09:23
So the, the great dream of Apple as I understand it is just that,
00:09:27
you know, we get to a point where a lot of these small
00:09:29
models can be run local, local and that they have these great
00:09:34
M1 and M2 whatever generation we're on now chips.
00:09:38
And that even though they're not GP, they're sort of combined
00:09:41
chips, right. So there is some graphical
00:09:44
component, right. And so that those chips would be
00:09:47
sort of pretty capable at handling sort of small local
00:09:50
models? Totally.
00:09:52
And I mean. Apple's argument would be we're
00:09:54
using transformer based models in photo recognition stuff,
00:09:57
we're using it in speech recognition stuff, we're using
00:09:59
it in auto correct. They just put it in the iPhone,
00:10:02
right. They're like their their take is
00:10:04
always like well we will do it when there's a real customer
00:10:06
need and this tool can really provide value.
00:10:09
So I think like Jury is still out there.
00:10:10
I also think with Apple Vision Pro it'll be really interesting
00:10:13
to see what kind of generative AI and and also just you know
00:10:18
the AI driven tool set that launches with Apple Vision Pro.
00:10:20
So I think like. It's a really interesting
00:10:23
question as to whether or not like this is whole AI revolution
00:10:26
is just going to entrench the power of the big 5 tech
00:10:28
companies because I think like they look pretty good right now.
00:10:32
You didn't. You didn't mention Amazon,
00:10:33
right? I I said Amazon was like cozying
00:10:36
up to anthropic. And also, I mean, you can
00:10:38
elaborate on the other stuff they're doing, which is quite
00:10:40
copious. On on Alexa specifically.
00:10:42
Well, and also AWS, right? Bedrock and everything, right?
00:10:45
Yeah, right. Well, it feels like Amazon, I
00:10:48
mean Microsoft is clearly ahead like they with Azure and AWS for
00:10:53
those two companies I think are the most important.
00:10:56
I get you guys really care about Alexa and stuff could happen.
00:10:59
But for the businesses today, like Amazon Web Services is so
00:11:03
essential. And like Microsoft can go to
00:11:06
people and say like Azure has a direct relationship with Open
00:11:10
AI, whereas Amazon, it feels like they've been pretty slow on
00:11:15
this. But like like you said, they
00:11:16
have Bedrock and they're trying to come up with these sort of
00:11:19
partnerships, but. Yeah.
00:11:21
Well, their strategy is they're going to put everyone else's
00:11:23
model on AWS, right. So they're going to be the
00:11:25
middleman, like they're going to sell you other access to other
00:11:28
people's models. So the anthropic models and I
00:11:30
think to the Llama models and to whoever other models they can
00:11:33
get on there, they're going to get everything on there.
00:11:34
And the only like maybe they will, maybe they won't end up
00:11:38
with open AIS models on there. I think that's an interesting
00:11:41
ongoing business discussion. But like, in the end, they're
00:11:43
going to try to be. You know the supermarket for,
00:11:46
for, for models in the cloud, right as they have been in the
00:11:49
the the last generation of all the different web tools they
00:11:52
sell. So like.
00:11:53
But maybe maybe Azure is not allowing Open AI to put that
00:11:58
model on AWS essentially right? Yeah.
00:12:01
So is Open AI so good that that will cripple Amazon?
00:12:06
But it is interesting that I imagine most developers that who
00:12:10
are using GPT models are. Doing that directly through the
00:12:14
open AIAPIS. Yeah, probably.
00:12:17
And there is, you know, there's this world now that was already
00:12:21
trending this way of big companies wanting to have like a
00:12:23
foot in Azure and AWS so that you can sort of get the best of
00:12:28
both worlds. But that would seem to be good
00:12:31
for Microsoft given that they were not the first place player.
00:12:34
So even saying OK this further creates this world where you
00:12:38
want hybrid cloud or you want to dual app clouds is is sort of
00:12:43
good for Microsoft. I I can't believe I'm talking
00:12:46
cloud computing on the podcast. I feel like even even even as
00:12:51
long as we're willing to go it's funny to be in that room.
00:12:55
OK. Facebook, I think part, you know
00:12:57
Facebook is releasing Llama which is sort of their, they're
00:13:01
sort of the king public company and open source models.
00:13:05
Their their their top AI person is Jan Lacun who's sort of he's
00:13:10
been at this forever and has had a lot of spicy and open takes.
00:13:15
I mean I guess the the core question is just like what do
00:13:18
you make of Facebook which is sort of a closed social network
00:13:23
being so pro like let's just give, let's give foundation
00:13:28
models away, make it easy, let's commoditize them.
00:13:31
Let's just like give away. What do you make of the Llama
00:13:33
strategy? Like what's?
00:13:34
What's the point of giving away Foundation models and making
00:13:38
that such a core piece of your approach?
00:13:41
I mean, I love it. I think it's good.
00:13:44
Yeah, I think it's great. I think it's really, it's really
00:13:48
probably primarily about developer relations and creating
00:13:56
goodwill in the open source and the developer communities.
00:14:00
That for them is more important overall to the like longevity of
00:14:06
their business than. Owning you know, or monetizing
00:14:12
these APIs, these models through APIs which would be a just an
00:14:17
entirely new business for them. So I don't know, that's that's
00:14:22
my take. I don't know.
00:14:23
Do you guys have a different perspective?
00:14:26
I just think that I think that there's basically 2 strategies
00:14:29
here with like open and closed source software, right?
00:14:31
One is like closed source where you keep it private.
00:14:34
You're like we have the best stuff.
00:14:36
And people will have to come to us even though they won't know
00:14:39
the sort of how the internal workings of the software are
00:14:41
because it's the best, right. And that's pretty much the open
00:14:43
the eye strategy, right, like and some of the other foundation
00:14:46
model companies like Anthropic whatever.
00:14:48
And then Facebook is like, OK, somebody already took the closed
00:14:51
source. We're the best strategy.
00:14:52
We're going to take the the, you know, opposite strategy which
00:14:55
is. We're going to open source all
00:14:57
these models, right? And we're going to hopefully
00:15:00
build developer communities around these models so that
00:15:02
people can make them better. They can, you know, fine tune
00:15:05
them. They can like, you know, help
00:15:07
optimize them over time. And then we will have like the,
00:15:12
you know, the. Thousands or, you know, 10s of
00:15:15
thousands of developers who are working on top of these models,
00:15:18
like build and improve them and build a network of of plugins
00:15:22
and, you know, optimize them and all this good stuff.
00:15:25
And so we're going to get value out of the developer community
00:15:27
and and therefore the models that our business Facebook is
00:15:30
built on top of will will get better and better that way,
00:15:32
right. And so I think it's like it may
00:15:34
not work, but I think it's better than just trying to.
00:15:38
It's differentiated. Yeah, It's differentiated,
00:15:40
right. Yeah.
00:15:41
And you know, there are lots of great startups that have been
00:15:44
built. You know where it's like we're
00:15:45
going to give away a foundation, we're going to give away, sorry,
00:15:48
an open source project and then you could build a business
00:15:50
around if it's super successful. Also, for Facebook, they
00:15:56
probably benefit a lot from hiring great AI talent in the
00:16:00
long run. And and if.
00:16:02
If that alone just gets them, you know marginally, you know
00:16:07
percentage. I know Reels is getting better.
00:16:09
You know, I I feel like the. The reels is good.
00:16:11
Reels is a good algorithm. When if Facebook and Google nail
00:16:15
AI generated ad advertisements, right, or add just even the
00:16:19
words in the advertisements, the copyright like that, you have to
00:16:23
imagine that's a huge tailwind for the business, right?
00:16:26
Like like. You know, personalized ads that
00:16:29
are created by AI in real time as you scroll to Instagram
00:16:32
feeder or the Facebook feed or whatever.
00:16:34
Like, you have to imagine that's.
00:16:36
Yeah, yeah, you have to imagine that's a crazy good business for
00:16:40
them, right? Or improvement to their business
00:16:42
and lowers the friction to actually buying the ads.
00:16:45
It's kind of interesting if you think of it that way, because
00:16:48
what you're saying Max is AI just improving and getting
00:16:52
better faster. Is, is is what's best for
00:16:56
Facebook overall, almost like if we had a breakthrough in, you
00:17:01
know, energy production of energy or clean energy or
00:17:04
something that would be great for Facebook too, probably
00:17:06
because they would, you know, have a lot, we could run their
00:17:09
data centers a lot cheaper and you know all these things,
00:17:12
right? So it's almost just like better
00:17:14
for them to encourage the advancement of this industry
00:17:18
faster cause a lot of the gains will accrue to them.
00:17:21
We we jumped into this with business strategy which I am
00:17:24
enjoying. But to me what's interesting
00:17:27
about Facebook's positioning is that you know Google, Microsoft
00:17:32
with Open AI to some degree they're playing the responsible
00:17:36
steward right. It's we are going to make sure
00:17:40
our language model behaves appropriately, doesn't get off
00:17:44
the rails. We're our brands are at stake
00:17:46
and we need to really guard them.
00:17:48
You know, Mustafa Suleiman at Inflection, who's going to be
00:17:51
speaking at the conference has sort of staked out and you know,
00:17:55
we're going to interrogate this more.
00:17:56
And so I don't want to caricature it before he speaks.
00:17:58
But like a somewhat open source, skeptical position, right?
00:18:02
It it undermines the idea that we need to really be thoughtful
00:18:06
and protective about who gets access to what and what these
00:18:10
systems can do. And so Facebook is playing sort
00:18:14
of AI, don't know flamethrower fire thrower type role here
00:18:19
where they're throwing fuel to the fire on open source in an
00:18:23
area where a lot of the other big tech companies would say the
00:18:26
thing we can do is like do this really well and then keep it
00:18:29
really closed off so that, you know, we have basically we're
00:18:32
almost as functioning as a state ourselves where we're really
00:18:36
being thoughtful. What do What do you take of
00:18:39
Facebook, the big company playing this role of
00:18:42
democratizing? I do worry, I guess for them if.
00:18:48
Someone uses Llama for nefarious purposes like.
00:18:51
Essentially, it hasn't. It's a brand, but.
00:18:53
Yeah, I think that could happen. I think that this could be like
00:18:57
a long running issue now for them going forward the next few
00:19:01
years. If for if, if, as we've talked
00:19:04
about in previous episodes, these models can be used for
00:19:07
nefarious, really is laying the groundwork for reporters here.
00:19:10
You're like, OK, people should be held accountable for what
00:19:12
their foundation models, their open source project does.
00:19:19
I don't. I don't.
00:19:20
Know if I I don't know if I personally agree that I'm saying
00:19:24
that will be the. That could be a narrative for
00:19:27
sure, and it just depends on how how much harm can be done with
00:19:32
open source models, and I think we don't know yet.
00:19:35
I mean. Facebook's like, look, our brand
00:19:38
for trust and safety is already so bad that there's nothing we
00:19:41
can do that could possibly make it worse.
00:19:43
So why don't we just go for it On the business front, they're
00:19:46
like Cambridge Analytica. Wait till they see what people
00:19:49
will do with this model. Well, Cambridge Analytica is a
00:19:52
great similar analogy because they thought they were.
00:19:57
At the time they were opening up the Thorngren graph and they
00:20:00
were, you know, giving data to, you know, other developers and
00:20:04
companies. And they were, and that was, you
00:20:06
know, encouraged almost at the time, right, by developers.
00:20:10
And it blew back on them entirely.
00:20:13
And it wasn't even good for their business to be doing that,
00:20:15
to be giving, leaking that data. So it is there is kind of a
00:20:19
similar thing occurring here potentially.
00:20:23
Facebook's willing to be wild. Well the narrative kind of can
00:20:26
change on you right that with at the beginning of social
00:20:30
networking and and technology this was encouraged that we
00:20:33
would want you know more more data sharing or something with
00:20:37
developers and that and then Fast forward 5-10 years and it's
00:20:41
not have we all staked out points of view.
00:20:44
I think we're all on the same page on do we believe in open
00:20:47
source like data bricks, super pro open source, like a lot of
00:20:50
the, you know, obviously hugging face clam at CV1, super pro,
00:20:56
open source. Like, I feel like generally
00:20:58
we've heard from a lot of pro open source people.
00:21:01
Do either of you have reservations?
00:21:04
Are you fully in the camp of like it's an arms race and just
00:21:08
like get the best stuff out and let good people sort of take
00:21:10
advantage of? It my take is that open source
00:21:15
right now is great for the community, for developers, it's
00:21:19
awesome to be able to. You know, edit weights or use
00:21:23
smaller models, run, run in or your own inference, all these
00:21:27
things fine tune, right? And that just might not be the
00:21:33
case going forward. I don't know, at some point
00:21:35
maybe I will change my mind and be like, no, the models are too
00:21:37
powerful. Like I don't want these open
00:21:40
source models in the hands of nefarious actors or you know,
00:21:45
China or something. So I I guess my take is it's
00:21:48
great right now. I don't know, but it's hard to
00:21:51
pull. You know, that's sort of the
00:21:53
once they they have the models and it's gone too far, it seems
00:21:56
hard, hard to unwind the clock or whatever.
00:22:01
Yeah, I guess I I just believe the nefarious actors in this
00:22:04
scenario are unstoppable. And as we discussed in The AI
00:22:09
Kills Us All episode that somebody's going to figure out
00:22:12
how to you know, some bad, bad actor is going to get control of
00:22:15
a a really powerful model regardless of whether or not it
00:22:18
was open source or not so. Yeah, I mean give the, so you're
00:22:22
you're very pro give the pro open source give the give the
00:22:25
fruits of the labor away to the community.
00:22:27
I mean why do open source? Why do we think?
00:22:30
Why do we think Open AI changed its tune so dramatically on
00:22:34
this? Question.
00:22:35
Awesome business strategy. Just like, just protect your
00:22:38
business. Good for the front runner,
00:22:40
right? Like like like regulate it.
00:22:43
Like let tell. Make sure you have to spend a
00:22:45
lot of money, like following rules like that's good for big
00:22:49
companies. Because once they realized how
00:22:50
good the stuff they were cooking was, why give it away?
00:22:54
I mean, open source is the strategy when you're not the
00:22:56
leader. Like not, you know, you never,
00:22:59
never give away the the high quality stuff, yeah.
00:23:02
Or or the strategy that makes sense for meta because they have
00:23:06
this these huge other businesses right like you would.
00:23:09
It doesn't really make sense. To give away models as a start
00:23:13
up I guess I I mean maybe maybe that's what stability is kind of
00:23:17
doing or mid journey but you said they're kind of like
00:23:21
becoming more close source. Imagine a world where like these
00:23:25
algorithms are near AGI like and then they are locked down.
00:23:31
You know that's the ultimate elite.
00:23:33
So like a a small group of companies gets to control like
00:23:37
this sort of Infinity resource like it seems that seems insane.
00:23:41
You know what I mean? I feel like that would become
00:23:44
sort of one of the great freedom questions of our time, that just
00:23:48
in any way restricting the set of people who have limited to
00:23:52
access to like a God like reasoning and information
00:23:55
system. I feel like we're we're
00:23:56
backtracking on Oh yeah kills us all episode but but I mean I do
00:24:01
think open source was key to Facebook and that's how we got
00:24:03
down this path. But reasonable that we've gone
00:24:06
far from that though you've been the one that makes the point
00:24:10
that like often these discussions undercount the AGI
00:24:14
potential and how disruptive that would be to any of the
00:24:16
other arguments that we're. Having I total, I totally agree.
00:24:19
And that's where I would question your assumption that
00:24:23
everyone should have AGI. Like Without knowing how
00:24:25
dangerous that could be for the world, that seems like pretty
00:24:29
presumptuous to just assume that that it should not be regulated
00:24:33
by a handful of companies. Yeah, I mean it's a lot just.
00:24:37
Give away the God like powers. Everyone deserves God like
00:24:41
powers. I mean, everyone deserves what
00:24:42
can go around. Yeah, I mean they they
00:24:46
presumably they will have some counter, you know
00:24:50
countervailing. Presumably, sure, yeah.
00:24:52
I guess if that's the case, you know, but, oh, I guess here's an
00:24:57
easy way to make sure we haven't missed anybody.
00:24:59
Let's Let's do a collective ranking of who we think benefits
00:25:03
the most from AI in terms of market cap movements experienced
00:25:09
and to be experienced. This is I'm putting, I'm putting
00:25:13
in no we're doing it together. I'm putting.
00:25:15
I think NVIDIA should be our number one.
00:25:18
I mean, Microsoft would be the close one there, I think.
00:25:21
I mean, Microsoft stock isn't actually up anywhere nearly.
00:25:25
So, so NVIDIA then probably, I mean to your point I guess, but
00:25:29
I guess the question. What about Google?
00:25:30
I I would put Google pretty high here, Microsoft.
00:25:34
Depends on what time frame. 34% over the year and 187% / 5
00:25:40
years? Google you're putting Google.
00:25:45
I don't know. What to me?
00:25:47
What time rules in though like tech becoming more important and
00:25:51
Google being all over it is good for tech is good for Google.
00:25:54
But like Google's relative dominance, which was so powerful
00:26:00
in search, feels threatened by this, so I I just.
00:26:04
No, I think that's worth talking about because I just feel pretty
00:26:10
confident that that whole search impact of AI has like is like
00:26:17
very, very early and Google will have plenty of time to
00:26:22
potentially continue dominance with you know using LMS and and
00:26:27
AI and multimodal models in in in their search products, so.
00:26:33
They have, they have the the front page right that everyone
00:26:37
goes to for search and they have really powerful AI internally.
00:26:42
It just hasn't played out yet. But I don't.
00:26:45
I believe that they will get there.
00:26:47
All right. So I mean you're you want Google
00:26:50
first it just. Depends what we're talking
00:26:53
about. Like are we talking about which
00:26:55
of the biggest companies will have the highest market caps
00:26:58
from AII mean or just like relatively like you know?
00:27:03
Is their position improved by AI?
00:27:05
How much the percentage of their position is improved by AI?
00:27:10
That's what you're asking. Yeah, I mean NVIDIA.
00:27:16
Yeah, NVIDIA for sure. Microsoft second agree.
00:27:24
Facebook Third I would go what? Yeah, I don't.
00:27:27
I don't agree, because algorithms are core to what they
00:27:29
do in their discovery. Mechanism.
00:27:31
I think if you actually, if you oh, because of algorithms are
00:27:34
core to what they do, OK. I just think if you believe
00:27:37
there can be a 50% improvement in advertising at scale through
00:27:41
AI generation, then Facebook benefits more from that than
00:27:45
anyone except Google and and they might think more than
00:27:48
Google. I have just put Google before.
00:27:50
OK, I mean, that's fine. It's, but it's a totally.
00:27:52
Agree with Eric, but but. It's just about search versus
00:27:56
the Instagram feed. Which one gets more benefits out
00:27:59
of AI like, I think that's the real question here, like or
00:28:03
search versus the Instagram feed plus the Facebook feed, right?
00:28:06
And I guess it also has YouTube which is competing in like
00:28:09
shorts. Yeah.
00:28:11
But it's going to take a while to do video generation, I guess.
00:28:13
I don't know, like it's sort of about the latter of how quickly
00:28:17
generated personalized ads kind of get scaled up.
00:28:19
So it's. Yeah.
00:28:21
Facebook, you're mostly talking about their ad products and.
00:28:24
Ranking algorithms. But there's also an argument
00:28:27
around on the content generation piece itself that as we've
00:28:31
talked about before, there's, you know, ability to create more
00:28:35
compelling content and social media posts and I.
00:28:38
I really think the only way you can make the case for Google is
00:28:41
that you would say the ads on Google are going to get much
00:28:43
better, right? I mean like.
00:28:44
Well, what about search itself? You think search is going to get
00:28:47
so much better I. Think search itself is going to
00:28:50
get so dramatically better? Yes.
00:28:52
Like I think, I don't know, I mean time spent in search you
00:28:55
would think that go up if they're chat type products.
00:28:57
Well, well, there is an argument that time spent goes down,
00:29:01
right? Because you're just getting your
00:29:02
answer faster at an end. But it's you're not.
00:29:05
It's no longer a link. It's like less of a link system
00:29:09
and more like in the search, which is where Google ads are.
00:29:12
Well, it just depends. If it's more, it becomes more
00:29:15
like a ChatGPT model where you're chatting with Bard or
00:29:18
instead of searching. I guess that's where I'm coming
00:29:20
from. I I see me and myself doing
00:29:23
that. Like almost more than.
00:29:25
I searched Google today in the future, and I think they're the
00:29:27
best positioned to be that like true assistant that I use every
00:29:33
day to talk to and instead of searching by typing into AI mean
00:29:38
I was down on Google before. But like Google versus Facebook,
00:29:40
I'm getting heated up just in the sense that, like, if you
00:29:42
think of every product that Google offers like if Gmail was
00:29:46
amazing from AI and like the prompts were great and there was
00:29:49
writing like could become an even more insane product.
00:29:52
Obviously Like Google cloud services.
00:29:56
Wait, sorry, it's Google. Cloud platform, global cloud
00:29:59
platform is key. If they can improve their like
00:30:03
search, YouTube, like they're just so many pieces of their
00:30:06
business where they have audience and they could put this
00:30:10
to work. The only question we all have
00:30:13
probably is execution on all of these areas, but the opportunity
00:30:17
to me seems massive. I mean, I guess I'm just going
00:30:21
based on the fact that Google makes no money on anything
00:30:23
except search and the other stuff is great, but it's really
00:30:25
just a data vacuum for for search ads, like they really
00:30:28
don't make any money. I mean, they make one on YouTube
00:30:30
now, so that would be the other big thing.
00:30:31
But like search ads is the whole ball game if you're just purely
00:30:35
talking about the business of Google, pretty much.
00:30:37
And so you have to believe that search ads are going to get way
00:30:40
better and or that people are going to spend way more time in
00:30:43
search, which again is an execution question about like
00:30:45
Bard versus all the other, you know, large language models and
00:30:48
chat models. So like I buy the case, I just
00:30:50
don't think it's as clear cut as you guys do because I just think
00:30:52
Facebook like Instagram ads and Facebook ads getting better
00:30:55
would like to immediately dump right into the business.
00:30:57
Which and what you're saying is that there's more of an obvious
00:31:02
trajectory for. The Instagram ads and feed to
00:31:06
get better from AI than there is for search ads to get better
00:31:09
from AI. Yeah, yeah.
00:31:10
I mean, it's close. They're, you know, they both are
00:31:13
going to get a lot better. So, Yep.
00:31:15
But it's about which one's bigger and what's the bigger
00:31:17
impact, I guess, Yeah, notably, notably like, so yeah, we, we,
00:31:23
it's Facebook or I don't know, I Facebook wouldn't necessarily be
00:31:27
my next. Either.
00:31:28
Oh, really? Interesting.
00:31:29
I'm. I'm willing to.
00:31:30
I mean, I. Just feel I would bet it forth
00:31:32
at least, but what? What's next?
00:31:35
We have Apple and Amazon, you know, like are now towards the
00:31:38
bottom of our at least tech list.
00:31:40
Obviously tech being better in some ways lifts all boats, but I
00:31:44
don't know what do you make of those two and like what it means
00:31:49
for them. I would put Amazon next.
00:31:52
I think they have an obvious role to play with AWS and they
00:31:58
have made a lot of great partnerships there with
00:32:00
Anthropic already. They are probably working
00:32:03
internally on models themselves. They are the number one cloud
00:32:09
platform. They will benefit in their other
00:32:12
core businesses through AI potentially, you know Amazon.com
00:32:19
and ads there, which is a increasing part of their
00:32:21
business and not to ignore it, you're allowed to talk about it.
00:32:27
Yeah, Alexa. Which is critical to our
00:32:30
business. They have an end point in 30% of
00:32:36
US homes that is an AI assistant.
00:32:40
It historically hasn't been amazing at talking to you
00:32:44
beyond, you know, controlling smart home features and and
00:32:48
telling you the weather and playing volley games.
00:32:51
But it's it's completely possible that they will be able
00:32:56
to. AD LLMS and and AI experiences
00:33:00
as they've already demoed that would transform that device in a
00:33:04
way that would be you know essential would would would be a
00:33:08
game changer for the Alexa business.
00:33:10
So I think they're the furthest ahead in that area on like
00:33:15
potentially integrating AI with, you know, your average day
00:33:19
through a Voice Assistant. And Apple, I mean Apple, I mean
00:33:23
the argument Apple gets away with everything in that we can
00:33:27
put them in like basically last place and it's still like, well,
00:33:30
they move slowly and maybe they'll figure something out.
00:33:32
And like, I don't know, it's definitely their wheelhouse.
00:33:35
It's like it's not, it's a lot of data.
00:33:38
It's not about privacy. It's sort of not a device.
00:33:41
I mean, I guess, well, I mean, if we really want to criticize
00:33:44
Apple, we saw this report that, you know, maybe Sam Altman and
00:33:48
Johnny I've and Masa son Masayoshi, son of SoftBank are
00:33:54
getting the team together weird. The weirdest team headline you
00:33:57
could imagine to build a phone or some device that's sort of
00:34:02
got AI at its core, that that's a threat.
00:34:05
I mean, I could buy, you know, a next generation sort of AI first
00:34:09
device. I I I just think Apple's Moat
00:34:11
has always been how unbelievably difficult it is to build
00:34:14
hardware at scale, right? Especially like high margin
00:34:17
hardware. I mean, there's almost no other
00:34:19
large hardware businesses in the industry.
00:34:21
And even if Johnny is on board and Masa is on board, you still
00:34:26
need like Masa's got a great history.
00:34:29
You still need, yeah, you still need about three to five years
00:34:32
and 30 or $5 billion to build even like a small scale, high
00:34:35
quality hardware device. And just like I think that's
00:34:37
just a long way away and anywhere where you know Johnny
00:34:42
and and. Sam are playing like Tim Cook
00:34:45
will pay some attention to that and make sure that they're not
00:34:47
completely caught off guard, right.
00:34:49
You know, they all talk to the same people manufacturing the
00:34:51
stuff in China, so but would you?
00:34:52
Agree though that like Apple, while they're not threatened
00:34:56
that much, like relatively, they are probably lowly ranked here.
00:35:00
Yeah, I think. I I think.
00:35:01
They're the most, most protected on the downside in some case, in
00:35:04
some cases you could argue and then also they don't have huge
00:35:07
outside this is all bullshit. They're fine.
00:35:09
And if it's good, you know, it doesn't feel like they're going
00:35:12
to disrupt it, yeah. Right.
00:35:14
I would like to throw another company in the mix here.
00:35:17
What do you guys think about Tesla?
00:35:20
Ah, yeah. Self driving.
00:35:22
The old self driving and the general xai suite.
00:35:26
Now, you know, it's like he's somehow allowed to like, merge
00:35:31
AI research, like it seems like across all his companies.
00:35:36
I mean, I just think that I think I'd be much longer way MO
00:35:39
and cruise than than Tesla or whatever.
00:35:41
I mean like, I just think that. Tesla's fundamentally built
00:35:44
around a hardware business model where you sell cars and like if
00:35:47
you believe we're entering a world where all every car is
00:35:49
self driving and it can take you anywhere.
00:35:51
Like why wouldn't you be long? Like, yeah, Uber or Lyft or
00:35:54
Waymo or Cruise or somebody who's built around this model of
00:35:56
like providing cars as a service rather than cars as like a
00:35:59
jewelry object that sits in your garage 99% of the time.
00:36:02
Like, I just don't like. I think Tesla's a great product,
00:36:04
but is it an AI for self driving first product?
00:36:07
I always thought. The idea that it was gonna turn
00:36:09
into a self driving taxi fleet was like pretty absurd, but.
00:36:13
I think that's a good point. I think Elon, his hype or
00:36:19
marketing spin on this is that the AI that they've built to.
00:36:24
Successfully get self driving. Working at Tesla is so valuable
00:36:29
in future endeavors, right? Like robots, robotics.
00:36:32
But didn't they literally admit they gave up on all the AI
00:36:35
they've been building for? Until a year or two ago, they
00:36:38
had been using the last generation of deep learning
00:36:41
tools or whatever. And then at two years ago, there
00:36:42
was a. Like recently there was a story
00:36:44
that came out that basically said, yeah we had to redo the
00:36:46
whole thing once we realized like transformer models, large
00:36:49
language models are better, right.
00:36:50
So like they they've been saying that for like 7 years that like
00:36:53
we're way ahead and everyone on building a self driving models,
00:36:56
we have gazillion trillion miles of data blah, blah, blah.
00:36:59
Which is fine is true. But like they had to throw it
00:37:01
all in the trash can like 18 months ago.
00:37:03
So like I I don't find that like a super compelling like case
00:37:06
that they have some Moat around that sort of stuff.
00:37:08
But you know, who knows? I mean, they do have.
00:37:11
They don't have a lot of trust with like regulators.
00:37:13
Like, I don't. I feel like, you know, Waymo and
00:37:16
Cruz have just tried to be as responsible as possible.
00:37:21
You know, I mean, Waymo has been very conservative in deployment.
00:37:25
And you know, Cruz is like General Motors, like a company
00:37:29
that I feel like every American politician is rooting for,
00:37:32
right. I mean, Tesla, it seems like now
00:37:35
has been able to position itself as very close to the Republican
00:37:39
Party, but it feels like the administrative state and
00:37:42
Democrats are going to be, like, extremely skeptical of them just
00:37:46
throwing cars on the road. I mean, that's sort of cynical
00:37:51
even about, I guess, the politicians I like.
00:37:53
But I yeah, I think I agree with Max's point of technology.
00:37:58
I also think having a reputation for being a responsible, sane
00:38:04
actor is important when you're doing something as dangerous as
00:38:07
deploying the first self driving cars.
00:38:11
I think it's a fair, fair assessment.
00:38:12
I mean we'll see if they pivot the attack, they're going to be
00:38:14
a stock premium for it. I'm sure like I guess they pivot
00:38:18
the you couldn't you can't I don't know how you can ever
00:38:21
measure this because it's like well the stock can stay inflated
00:38:24
for like forever So if they if they pivot.
00:38:26
The entire company to robo taxis and they start selling $5 Tesla
00:38:30
rides all around America. I will totally change my tune on
00:38:32
this, but you pretty much have to embrace a disruptive business
00:38:35
model in the next three to five years, which is historically
00:38:38
been very, very difficult for any type of company because
00:38:41
you're kind of saying all this money we're making a day.
00:38:43
That's going to be 0 in five years and we're just going to
00:38:45
have to jump on the new thing and ride it out, which is really
00:38:48
tough. I to be clear, I don't think the
00:38:51
self driving deployment cycle will be anywhere near as fast as
00:38:55
that would suggest. So I don't think it's going to
00:38:57
destroy their like existing car business.
00:39:00
But I agree with the idea it's hard.
00:39:03
The the classic start up situation which you're
00:39:06
articulating is that like people don't disrupt their own good
00:39:09
businesses and so Tesla's in sort of the worst situation for
00:39:12
that. Just a counterpoint to that, do
00:39:16
we believe that other companies are going to be able to sell
00:39:20
fully self driving vehicles anytime soon?
00:39:23
It doesn't doesn't seem like it, at least besides the three we've
00:39:27
talked about. No, I I mean those other
00:39:28
companies seem to be only focused on the robo taxi market
00:39:32
as opposed to getting consumer vehicles that can fully self
00:39:37
drive themselves into the world like if.
00:39:40
Tesla is the only game in town there, and I'm not sure if
00:39:43
that's the case, but right now it seems to be the only one that
00:39:46
is going in that direction. That could end up being a very,
00:39:50
very compelling offering if the only car you can buy that will
00:39:55
drive itself anywhere is a test big if we'll see.
00:39:59
I feel like Waymo and crews are going to deploy city by city
00:40:02
geofence where they know they can do it with no deaths, you
00:40:06
know? Yeah.
00:40:08
Anyway. Yeah, are there any other, I
00:40:11
mean this is great. I I'm glad you brought up Tesla.
00:40:13
Like what Any more edge case companies we haven't considered.
00:40:18
I mean this is part of the problem I think right now and
00:40:19
why I think a data bricks IPO and the Mosaic deal will be
00:40:23
welcomed. It's like there's a public
00:40:26
market desperation for a way to bet on AI, which is why I think
00:40:31
NVIDIA has done so well because, you know, these big tech
00:40:36
companies are already very highly valued.
00:40:39
You know, the startups are private.
00:40:41
Cool, great. Thanks for coming on.
00:40:44
And now we're going to Chris Miller with Chip War, who's a
00:40:48
real genuine expert in the wonky world of chips.
00:40:53
And so we go deep in NVIDIA sort of their history and you know,
00:40:58
the, the big geopolitical question around the race with
00:41:03
China. So give it a listen.
00:41:09
Chris Miller, author of Chipwar. Thank you for coming on the
00:41:12
show. Thank you for having me.
00:41:14
I feel like so much of what sort of the regular person or even
00:41:19
you know, I don't know, the average startup founder, venture
00:41:22
capitalist in Silicon Valley thinks of this artificial
00:41:26
intelligence phenomenon starts and sometimes ends with chat GPE
00:41:30
3 and it's certainly it's very much the like.
00:41:33
What bots can we play around with?
00:41:34
Like how am I talking to some sort of AI but you know in the
00:41:39
background doing reporting on some of these conversations
00:41:43
NVIDIA and like H1 hundreds and like weird chips come up over
00:41:48
and over again and sort of people's ability to get access
00:41:51
to those chips. So I really wanted to have you
00:41:53
on just to sort of like interrogate that and understand
00:41:57
sort of the technology behind it.
00:41:59
Can we just start off like, when when does NVIDIA start to matter
00:42:04
as a company? Or just give us like, the brief
00:42:06
history of this company, because it was like a weird, weird sort
00:42:09
of gaming chip making company, right for a long time.
00:42:13
Yeah, that that's right. NVIDIA was founded over 3
00:42:15
decades ago, but its earliest couple of years was all in
00:42:20
graphics, and graphics was most important for games, computer
00:42:24
games, video games. Which, you know, think back two
00:42:28
decades, they had the most complex graphic demands, figures
00:42:30
moving across the scene, pictures changing very rapidly.
00:42:34
And there were special types of chips that were produced by
00:42:37
companies like NVIDIA for processing graphics.
00:42:39
And if you think back, not so long ago, people would buy
00:42:43
specific computers to have the best graphics capability.
00:42:46
Now we sort of take it for granted, but for a long time
00:42:49
that was a real differentiating factor.
00:42:50
Can you explain sort of the difference between AGPUA
00:42:55
Graphics Processing Unit and ACPU?
00:42:59
So it's ACPU, which is the the workhorse of traditional
00:43:01
computing and computers and data centers.
00:43:04
They're very good at doing many different types of things, but
00:43:07
they do every computation serially, 1 after the other.
00:43:11
And for most use cases that's what you want because you're
00:43:14
undertaking different types of calculations.
00:43:16
But for for training AI systems, you want to undertake the same
00:43:21
type of calculation repeatedly. And so parallel processing,
00:43:24
which is what a GPU does, is capable of doing multiple things
00:43:28
in parallel at once, which is why they're vastly faster for
00:43:32
this type of calculation than a CPU is.
00:43:35
But why? Why does AI?
00:43:37
Why do AI tests need to be parallel while like a normal
00:43:42
computer task is sequenced? Well, they don't have to be.
00:43:45
It's just really inefficient, right?
00:43:49
You know, I think the key trend, there's been great research
00:43:51
looking at the amount of data on which cutting edge AI systems
00:43:55
are trained and and what you find, if you look at this for
00:43:57
the last like 10 years or so and chart it, the amount of data
00:44:00
used in cutting edge systems is doubling every six to nine
00:44:04
months. There's extraordinary demands
00:44:06
for putting as much data as possible into training.
00:44:10
And in order to make sense of all this data, to train your
00:44:13
model to learn from the data, it has to undertake lots of
00:44:16
calculations. And so the the computing demands
00:44:19
of training have shot upwards not just exponentially, but
00:44:23
exponentially, at at a rate that's faster than Moore's Law.
00:44:27
And that's why having the right ship for training has become
00:44:30
very important. Did NVIDIA see this coming?
00:44:33
Like, you know, I their stock is up year to date.
00:44:37
I think it's some like 200%. I I feel like Jensen Huang, the
00:44:41
CEO, you hear about them all the time.
00:44:42
Like was this a happy accident? It's like we're we're making
00:44:46
great video games and all of a sudden people want to buy our
00:44:48
chips for the hottest thing in the world or did they have some
00:44:52
foresight here? How much was this a plan on the
00:44:55
part of NVIDIA? You know, it certainly was a a
00:44:57
bit of an accident and if you talk to the company they'll
00:45:00
they'll they'll literally admit that when they founded NVIDIA.
00:45:03
They had no idea that AI would be an application.
00:45:06
But over a decade ago, they began to realize that there were
00:45:10
a a bunch of PhD students and researchers at places like
00:45:12
Stanford and Berkeley who were using their chips for things
00:45:15
other than gaming for complex calculations.
00:45:20
And that NVIDIA realized that if they tried to build out chips
00:45:24
around this application and no less importantly, build out a
00:45:28
software ecosystem around it, that they could.
00:45:32
Provide the chips that AI would require.
00:45:33
And so they've been investing very heavily for now, over a
00:45:36
decade. And for a long time that
00:45:38
investment seemed actually quite foolish because most of their
00:45:42
money was still in in in graphics and in gaming.
00:45:46
It's only quite recently that a data center in AI has become
00:45:49
their their primary. Well, you just, people talk
00:45:52
about artificial intelligence as a way to boost their stock price
00:45:56
or to, you know, seem like a more futuristic company than you
00:45:58
are for NVIDIA. It's like, OK, you're a gaming
00:46:01
company guy. Like, chill out and now you know
00:46:04
it's been totally validated. That sort of all this talk of
00:46:07
artificial intelligence wasn't total wild hype, hypemanship or
00:46:13
whatever. Well, and in between though,
00:46:14
they were also a a Bitcoin mining.
00:46:16
Yeah, I know, yeah, exactly. They did have the crypto.
00:46:19
They're definitely good at seizing the moment, but they
00:46:22
were they were being used for crypto mining, right?
00:46:24
Is that what it was? That's right.
00:46:27
And like, yeah, how do you see Jensen the CEO, like he he's one
00:46:30
of the Co founders writer. What's sort of his role at the
00:46:34
founding and how do you see him as sort of ACEO today?
00:46:39
He was one of the three Co founders of a video they they
00:46:41
met in a Denny's in San Jose to devise the business plan that in
00:46:46
the the 1990s and and Jensen I I think deserves a lot of credit
00:46:51
for realizing. A decade ago, that AI could be a
00:46:56
real growth driver. At the time, people thought it
00:46:58
was a really irresponsible thing to do plow all this money into
00:47:01
building a software ecosystem around chips.
00:47:03
No one else did that. Most companies just made chips.
00:47:07
And if you go back to the debate in 20-15 around that time
00:47:13
period, you can find lots of equity analysts saying, you
00:47:15
know, what's this company doing spending all this money on
00:47:17
software rather than selling chips?
00:47:18
It's a waste of money And and worse, they were giving the
00:47:21
software away for free. But it turned out to be a
00:47:24
brilliant move because everyone started using the the CUDA
00:47:27
ecosystem, which is is the the software layer on top of NVIDIA
00:47:32
chips. And that's put it at the center
00:47:35
of the the AI world. A key piece of the chip world is
00:47:38
this sort of like are you designing chips, are you the
00:47:42
fab? Can you just sort of explain
00:47:44
that and where NVIDIA sort of fits in?
00:47:46
Like, are they? Are they actually we say build
00:47:49
the chips. Are they like building the
00:47:50
chips? Yeah that that's a that's a key
00:47:53
distinction. They're not actually building
00:47:54
the chips. They only do chip design from
00:47:57
day one. They were, they were what's
00:47:58
called a fabulous chip company. They didn't have a fab.
00:48:01
They only did the design and and chip design is in many ways like
00:48:06
a a software design type of business.
00:48:08
You you basically write the code of the design and then you
00:48:11
e-mail it off to the the factory or a fab and semiconductor
00:48:15
parlance and and the fab actually does the manufacturing
00:48:18
and so in the case of NVIDIA. Their most important
00:48:23
manufacturing partner has been TSMC, the Taiwan Semiconductor
00:48:26
Manufacturing Company, which manufacturers most of their
00:48:29
chips. And so today for all of Nvidia's
00:48:31
key AI training chips, they're all manufactured by one company
00:48:36
in Taiwan. And do they they make all
00:48:39
basically all NVIDIA chips or? They make, they currently make
00:48:43
all of the key AI chips that NVIDIA produces and they.
00:48:47
Been the most important producer of video chips from day one, and
00:48:51
is NVIDIA like, how far ahead is NVIDIA right now in this sort of
00:48:55
AI chip? Are they singular?
00:48:58
Like, yeah, where where are they today?
00:49:01
Well, in terms of market share, Nvidia's far, far ahead.
00:49:04
So it's estimated that up to 90% of cutting edge AI systems are
00:49:09
trained on NVIDIA chips. And you can debate what's the
00:49:12
definition of cutting edge, but Nvidia's market position is.
00:49:15
That is extraordinary right now and that's reflected in in in
00:49:19
their stock price. I I think if you asked are they
00:49:21
technologically ahead of of competitors, whether AMD, which
00:49:25
is another chip maker or Google, which has a chip called the TPU,
00:49:29
which is also used for AI training, that that's a harder
00:49:31
question to answer. Exactly.
00:49:34
Yeah, exactly, exactly. And that's a hard question to
00:49:37
answer because there's both, you know, technical specifications
00:49:40
you could look at, but there's also ease of use and so if
00:49:43
you've built a system. Trains it on NVIDIA if you
00:49:45
yourself as a engineer have gotten comfortable with
00:49:49
invidious products and with CUDA.
00:49:51
There's also a a a switching cost of moving to somebody else,
00:49:54
and so Nvidia's both got their own capabilities.
00:49:56
They also got the first mover advantage.
00:49:59
Everyone's used to their products, but but they're really
00:50:02
scarce right now, right? I mean, my sense talking to
00:50:05
startups and investors, I mean, you hear Nvidia's like in every
00:50:09
venture round now because as part of the funding, they can
00:50:12
sort of promise people their chips, which then they make
00:50:15
money back from the startups. It's like a great world.
00:50:17
They're playing kingmaker. Like what?
00:50:20
What's your sense of how scarce their artificial intelligence
00:50:24
chips are? Right now they're scarce.
00:50:28
I hear the same anecdotes that you're hearing and NVIDIA and
00:50:31
TSMC, the manufacturer partner have have both said it's going
00:50:35
to be a year or so until supply normalizes.
00:50:39
And the problem is actually not that TSMC can't produce enough
00:50:42
chips, but they can't package them because they use a a very
00:50:46
complex, a new packaging technology to put the chips
00:50:49
right next to the specialized. I assume you don't mean in a
00:50:52
box. Yeah, you're like, you place the
00:50:54
chip on something, right? Yeah.
00:50:56
That's right, yeah. It's a Traditionally chips were
00:50:59
packaged in. That'd be funny though, if it's
00:51:00
like. Oh, the Apple cardboard box is
00:51:02
like holding them up at the end anyway, so go ahead.
00:51:06
Yeah. So, so packaging traditionally
00:51:08
was a really boring part of the chip business.
00:51:10
You take a PIC a chip, you put it in a a plastic or a ceramic
00:51:13
package and that was that was done.
00:51:14
But now it's it's increasingly important because you need to
00:51:18
get your your processor and your memory as close as possible with
00:51:21
as fast and interconnect as possible.
00:51:24
And so companies like NVIDIA are leaning more heavily on having
00:51:29
the right packaging capabilities to provide the most advanced.
00:51:34
Outcomes. And so it's actually TSM, CS
00:51:36
packaging capabilities that are the limiting factor right now
00:51:39
for producing more NVIDIA GPUs. And and what is an H100 or like
00:51:45
what what, what are, how different are these chips from
00:51:48
the the GPUs that everybody got excited about in the 1st place?
00:51:51
Like are they server chips Are explaining, you know, what is an
00:51:55
H100? But it's just the newest version
00:51:58
of of AGPU that NVIDIA offers specifically for AI training
00:52:02
and. The reason people get excited
00:52:04
about new versions of chips is because the rate of change is so
00:52:07
fast that a new version is always not just incrementally
00:52:10
better, but a lot better than the old version.
00:52:14
You know, it's it's it's not like the iPhone 15, which is a
00:52:17
bit better than the iPhone 14. But we've stopped noticing and
00:52:20
I'm like, this is not enough necessarily to get me from 13
00:52:25
Pro or whatever. Yeah, yeah.
00:52:27
The chips are different because of the the Moore's Law dynamic,
00:52:31
which is, you know, says you roughly double your processing
00:52:34
power every two years. And that's not a perfect, a
00:52:37
perfect metric, but ballpark that tells you just how rapidly
00:52:40
the technology improves. And the H100 is the most
00:52:43
advanced chip that NVIDIA offers.
00:52:46
So yeah, to put Moore's Law is still alive today.
00:52:50
Moores law is actually not a law.
00:52:52
It's a prediction that was set out by Gordon Moore, who founded
00:52:55
Intel in 1965. He predicted that the number of
00:52:59
transistors per chip would double every year or two, which
00:53:03
means that the computing power of chips would double every year
00:53:06
or two. And that's been and he said it
00:53:09
would be a decade, but then it's been true for a long time,
00:53:11
right? That's right, I'm just stealing
00:53:14
from your book. Of course it makes me sound
00:53:17
smarter when you get to read the book in advance, so you know all
00:53:20
the answers. Anyway, go ahead.
00:53:22
But yeah, that's right. So Gordon Moore thought, would
00:53:24
last for a decade, and here we are half a century later.
00:53:27
And Moore's law is is still basically holding up the the
00:53:31
rate of change increases and decreases at certain times.
00:53:34
But it's basically true that if you wait 2 years, a new chip
00:53:38
will have twice as many transistors and therefore twice
00:53:40
as much computing power. I mean, this is sort of a hard
00:53:44
question, so I don't know if you there's an easy answer, but
00:53:48
yeah, how much credit would you give then the improvement of
00:53:52
GPUs for what's going on in artificial intelligence right
00:53:56
now relative to, you know, attention is all you need or you
00:54:00
know, the papers in terms of the approaches to actually use these
00:54:04
chips to to run, run the software?
00:54:09
Well, you know, I think it's certainly it's the case that
00:54:11
that there's there's multiple factors driving innovation.
00:54:14
But I just the way I like to think about it is, you know
00:54:17
what's improved most over the last decade?
00:54:19
Is it the case that that software engineers are 16 times
00:54:23
smarter or is it the case that? Algorithms are 16 times better
00:54:28
or is it the case that because of Moore's Law 2 to the fifth,
00:54:32
we've got, you know, a vast increase in the number of of
00:54:36
transistors? You know, I think it's it's
00:54:38
primarily computing power that's that's driving us.
00:54:41
That's what you want. The chips guy comes on, he says
00:54:44
it's the chips. Yeah.
00:54:46
A key theme, you know, obviously in the book and sort of for
00:54:49
anyone thinking about the situation is, you know, the
00:54:52
rivalry with China and and sort of the the delicacy of the
00:54:59
global supply chains and you know, the, there's the chip
00:55:03
rivalry with China. But also in artificial
00:55:05
intelligence, you know, if we're worried that this is the
00:55:08
potential sort of path to some sort of, you know, generalized
00:55:12
intelligence, there's also the real question of whether the US
00:55:16
gets there first. Yeah.
00:55:20
So I mean, there's so much to hit at.
00:55:22
I mean, I guess the first question for me is, is the
00:55:25
United States getting most of Nvidia's chips or like how much?
00:55:29
How many of those chips are going to China?
00:55:31
Is China getting access to those right now while they're scarce?
00:55:36
Well, the the US last year imposed new rules that said
00:55:41
NVIDIA can't transfer its most advanced GPU to China.
00:55:44
And not just NVIDIA. Any company producing chips
00:55:47
above a certain threshold can't transfer them to China, so H1
00:55:51
hundreds are illegal to. To transfer to China, China's
00:55:55
not getting any of those. But but NVIDIA has design chips
00:55:58
that go basically right up to the threshold called A8
00:56:01
hundreds. And Chinese firms are reportedly
00:56:03
buying very large volumes of those.
00:56:06
So China's getting a lot of GPUs, but they're less advanced
00:56:09
than what AUS firm can buy. What would their next best
00:56:12
option be? Well, the next best option is
00:56:15
Nvidia's second best chip. Oh, man, OK, right.
00:56:20
And China tried to build its own fab, right?
00:56:24
A competitor of TSMC. What's, what's the status of
00:56:28
that and how much? You know it's fun, especially in
00:56:31
the US and given the market returns, like to talk about
00:56:33
NVIDIA, but like how much does this all hang on TSMC?
00:56:38
So China has been investing very heavily in its own chip
00:56:40
industry. It's got both companies doing
00:56:42
the design side and the manufacturing side.
00:56:46
There there are a number of Chinese GPU designers that seem
00:56:50
to have pretty competitive products, although even in China
00:56:54
Chinese firms prefer to buy invidious chips, so they're
00:56:57
they're not as good it seems. But the big challenge in China
00:57:01
is the manufacturing side because TSMC, the Taiwanese
00:57:04
firm, has been around five years ahead of the leading Chinese
00:57:09
firm SMEC. For a very long time, for at
00:57:12
least a decade, there's been a five year gap.
00:57:13
So both companies regular regularly improve the
00:57:16
manufacturing processes, but the rate of change in both is
00:57:20
roughly the same. So TSMC is always half a decade
00:57:23
ahead, and that holds to up the Today.
00:57:25
And what about the United States?
00:57:26
Like it hadn't Intel been trying to become a fab?
00:57:29
Like what's what's the state of our ability to actually
00:57:32
manufacture these chips if you know, there's a war in Taiwan,
00:57:35
God forbid? Well, yeah, that that's that's
00:57:38
the dilemma. So Intel does make its own
00:57:40
chips, but around five years ago it faced some severe problems
00:57:45
with its manufacturing operations.
00:57:47
And so for its most cutting edge chips today, Intel now turns to
00:57:51
TSMC for the manufacturing. Now Intel right now is is trying
00:57:55
to reformulate its manufacturing processes and it hopes that by
00:58:00
they're saying by 2025 they're going to be producing.
00:58:04
Have their most advanced chips back in house again, and that
00:58:06
those chips will be as capable as a TSMC made chip.
00:58:09
But right now, Intel's most advanced chips are actually
00:58:12
manufactured by TSMC. And then there's like a key, the
00:58:16
people who make the machines that help T that allow TSMC to
00:58:22
do what they do, There's like one company of those or what's
00:58:24
that company? Yeah, that that's right.
00:58:27
So TSMC, they know how to use the machines to make chips, but
00:58:30
the machines themselves are produced by a handful of other
00:58:33
companies, a couple in California, a couple in Japan,
00:58:36
and then one in the Netherlands called ASML, which produces the
00:58:39
most complex of these chip making tools.
00:58:43
So is quantity. Everything here with the H1?
00:58:48
Hundreds like, I mean, you know, I'm at, you know, Microsoft has
00:58:51
this cloud computing offering. You know, Google has its, Amazon
00:58:55
obviously has Amazon Web Services.
00:58:57
How much are those services trying to stockpile AI chips?
00:59:01
And you know, then these startups, whether it's, you
00:59:05
know, anthropic or open AI, how much is their game?
00:59:08
Just like we're going to win if we assemble the most H1
00:59:11
hundreds, Is quantity sort of the game right now in terms of
00:59:15
getting access to these chips? So long as there are shortages,
00:59:19
I think Quantity's a key part of the game.
00:59:22
And and what you see is that NVIDIA now realizes it's got a
00:59:25
lot of influence over the future of the cloud computing market.
00:59:29
And so NVIDIA has been, I I think, pretty actively trying to
00:59:33
build up other cloud computing firms by giving them exactly
00:59:39
exactly giving them access to GPUs and NVIDIA.
00:59:43
I think quite rationally wants a more fragmented cloud computing
00:59:46
market because that's a market in which it has more market
00:59:49
power relative to Microsoft, to Amazon and to Google.
00:59:54
So so we alluded to this, but sort of can you walk me through
00:59:57
the tech giants in terms of credible competitors to NVIDIA?
01:00:03
I mean we talked, I guess start with Google, it sounds like
01:00:05
maybe they're in the lead. I want to know about Intel.
01:00:07
And then is there anybody else who's who's relevant or even
01:00:10
close? Yeah.
01:00:12
So Google has a a in house designed chip.
01:00:16
They've got a big design team that has designed a chip called
01:00:18
ATPU which stands for tensor processing unit but fairly
01:00:22
similar to AGPU that is used for for AI applications and in
01:00:28
Google has a vast cloud computing business as well as
01:00:31
its own, its its own vast data center.
01:00:34
So lots of experience running very complex computing
01:00:39
operations but empirically we know that companies prefer to
01:00:44
use NVIDIA chips at least they do right now.
01:00:46
You know that could change as Google's chips get better as
01:00:50
scarcity drives up the price of GPUs.
01:00:53
But that hasn't been the the trend, the preference has been
01:00:55
for NVIDIA. Then AMD is the other company
01:00:57
that that is a competitor that in producing GPUs it's a a chip
01:01:02
design firm. They also manufacture with TSMC
01:01:06
that in. They're an American firm.
01:01:08
That's right. Based in Texas and and they
01:01:12
manufacture fairly competitive GPUs.
01:01:14
But the key difference with between their chips and Nvidia's
01:01:18
is that NVIDIA has the the ecosystem around it, the CUDA
01:01:21
ecosystem. And so once you're bought into
01:01:24
that ecosystem, right now the switching costs are substantial
01:01:28
enough where you just prefer to stick with NVIDIA if you can get
01:01:30
access to enough chips. Is Intel relevant at all?
01:01:34
Intel is is trying hard to produce competitive chips and I
01:01:39
think over the next couple of years we'll see it roll out a
01:01:41
series of new products that are designed to compete.
01:01:44
But right now it's a small player.
01:01:46
In my understanding like in cloud computing, like part of
01:01:49
the beauty of like an AWS is that you can sort of you know,
01:01:53
it's flexible, it's elastic, we can sort of have some people do
01:01:56
the computing, then you sell it to another customer at a
01:01:59
different time. Why hasn't that been the case
01:02:02
with this artificial intelligence computing, I mean?
01:02:05
I hear people talk about it like you're you're running full bore,
01:02:08
like these data centers are super hot.
01:02:10
Like what's what's so different about AI processing?
01:02:14
I think the the, the key difference is that the demand
01:02:18
for compute power in AI training is so large that there's just a
01:02:23
deficit of it. There's just a deficit.
01:02:25
And so the the systems that firms like Open AI are trying to
01:02:28
train are so vast that they they can't share their infrastructure
01:02:34
because they need all of their infrastructure and they need
01:02:36
even more of it than they actually can get access to.
01:02:38
And so the cloud computing business has been a business, as
01:02:42
you say, about engineering more efficient systems via sharing.
01:02:46
But in AI training, no one wants to share because everyone needs
01:02:50
more compute than they can actually get access to.
01:02:53
Right. And then there is also this sort
01:02:54
of perceived almost like 0 sum, like you want to deny your
01:02:58
competitor access. One company we haven't talked
01:03:02
about, you know, that we talked about in the laptop world for
01:03:07
chips is Apple. Like what's what's the status of
01:03:10
Apple on this? Sometimes people talk about
01:03:12
Apple and artificial intelligence on like local
01:03:15
processing. Is that a possibility or how do
01:03:17
you see Apple as being relevant or not in this race?
01:03:23
Yeah, you know, thus far we've been talking about AI training
01:03:26
and and AI training happens almost exclusively in big data
01:03:29
centers. But there's a big question about
01:03:31
what inference will look like in the future.
01:03:34
Some people think that inference will mostly happen in data
01:03:37
centers, that you're gonna ask a question of ChatGPT, you'll send
01:03:41
it back the Internet to the data center, tips to the data center.
01:03:44
We'll think about it and give you an answer back.
01:03:46
And that is efficient in some ways.
01:03:48
You have all the compute, all the chips in one big data
01:03:50
center. And so you can start to set up
01:03:52
some of the sharing and efficiencies that that we just
01:03:54
discussed. But the downside is you have
01:03:56
latency issues if you're sending all the data back and forth and
01:03:58
there's costs associated with moving data.
01:04:00
And so an alternative paradigm is doing more of your inference
01:04:04
on the edge of networks in your phone, in your car, for
01:04:07
autonomous driving systems, in in in your PC.
01:04:11
And that's where companies like Apple could start to play a much
01:04:14
bigger role. Right now, I think the the
01:04:17
market for inference on the edge is still very much in flux.
01:04:21
We're in the process of seeing many new types of chips rolled
01:04:25
out precisely for that purpose. So, so it's just sort of depends
01:04:29
on where it's one of these like sort of personal computer versus
01:04:34
server level questions back in the 90s where it's like OK, is
01:04:38
computing going to continue to happen?
01:04:40
In the cloud? Or is there some model where
01:04:42
local computers are going to do and you're saying that if it is
01:04:45
local, then maybe Apple's in a better position because they
01:04:48
have these great chips for for individual computers.
01:04:51
Is is that right? Yeah, that that, that's right.
01:04:54
And there's I think a pretty straightforward cost equation
01:04:56
of, you know, if the cost of data transfer is high, you'll
01:05:01
try to do as much as your processing can on the edge.
01:05:03
If the cost of the transfer is low and the benefits of having
01:05:07
big centralized facilities are are substantial in terms of the
01:05:10
efficiencies you can reap, then you'll have a much more
01:05:12
centralized inference process. And it it might vary even
01:05:16
between different use cases. In a car, for example, if you've
01:05:20
got a pretty autonomous car, your willingness to to depend on
01:05:25
a data center to tell you to turn left or to turn right, it's
01:05:28
going to be pretty limited. I would, I would, I would
01:05:30
hypothesize. Whereas for your phone maybe.
01:05:32
You want to make sure no matter what you can get the answer but
01:05:35
you know if it's for mid journey I can wait till exactly you
01:05:41
know, I mean we we talked about NVIDIA earlier on in terms of oh
01:05:46
there was a brief period where they were essential to crypto
01:05:49
mining and I and one of the key storylines I think think for the
01:05:54
non crypto world that was just watching it.
01:05:56
Was all the energy consumption around those crypto mines is?
01:06:02
Is that true for AII haven't heard that same environmental
01:06:05
story get picked up in this case.
01:06:08
But are are we destroying the environment and our pursuit of
01:06:12
artificial intelligence? Why, I think we we haven't heard
01:06:16
much about it just because there's such an extraordinary
01:06:18
demand for computing that companies haven't gotten around
01:06:23
to thinking about the consequences.
01:06:24
But the short answer is, is yes, it's extraordinarily energy
01:06:28
intensive and and we're going to get more efficient at it as as
01:06:32
time passes. But right now there are, you
01:06:36
know, such demands for energy in data centers that there are some
01:06:39
places in the US where it's difficult to build a data center
01:06:42
because there's no spare electricity capacity for the
01:06:45
data center. It's it's already a gating
01:06:47
factor. Have you seen any?
01:06:49
Yes, it's about energy use or anybody trying.
01:06:53
Well it's it's, it's tricky because everything depends on
01:06:56
how much more efficient you think things will get over time.
01:06:59
So you've got to assume we're going to be a lot more efficient
01:07:01
making the same compute available at lower energy demand
01:07:05
in 10 years time. But at what curve it's it's very
01:07:08
difficult to say. Yeah.
01:07:10
And to be clear, I think what what is the point of energy if
01:07:14
not to sort of push the cutting edge of like computing?
01:07:17
It seems like a very valuable use to me, but it yeah, it'd be
01:07:21
interesting. I mean, you obviously want to
01:07:23
make sure companies aren't being sort of cavalier about it.
01:07:26
And yeah, well, there's certainly not cavalier because
01:07:30
if you look at the cost of running a data center, 'cause I
01:07:32
don't know what it is, yeah, critical cost.
01:07:35
Well, I mean, I don't know if you're Open AI, you raised $10
01:07:38
billion, it's you can become sort of irresponsible.
01:07:42
You know there isn't the same market pressure if these
01:07:45
companies are able to raise what are basically like research
01:07:47
grants to to figure this all out.
01:07:51
Can you, I mean, we sort of jumped into artificial
01:07:53
intelligence, but I mean can the broad theme of the book, the
01:07:57
idea that sort of that we often think of global infrastructure
01:08:02
and the sort of weaknesses and vulnerabilities in terms of like
01:08:05
oil, right. I mean it's like oh, did we go
01:08:07
to war in Iraq over oil? What what is sort of the real
01:08:12
sort of what are the weaknesses with chips and what what why do
01:08:16
you think sort of this chip's arms race is?
01:08:19
So important when we're thinking about sort of geopolitics.
01:08:24
Well, the reason I, I, I wrote chipwork was when I realized
01:08:27
first, that we're surrounded by thousands and thousands of chips
01:08:31
over the course of our daily lives, Not just our phones and
01:08:34
our PCs, but it's cars and dishwashers and coffee makers
01:08:37
too. Second, when I began to
01:08:39
understand how critical they are for AI.
01:08:42
And then third, when I realized how concentrated the production
01:08:45
is in just a tiny number of countries and really just a
01:08:47
handful of companies, Taiwan being the the, the, the, the
01:08:52
most surprising example of that. And so when you look at the the
01:08:56
US, China race to develop more advanced AI systems and in tech
01:09:01
in general, then realize that both China and the US are
01:09:04
dependent to a shocking degree on ships made in Taiwan to power
01:09:07
their most advanced AI systems. It's an extraordinary situation
01:09:11
that we've all found ourselves in and it's a very dangerous one
01:09:14
given the concentration and given the dependency.
01:09:18
What would be the consequences of China having a chip edge
01:09:22
broadly and an artificial intelligence?
01:09:25
Well, just like we're preventing China from accessing our most
01:09:29
advanced ships, I think one has to assume that any country with
01:09:32
that position would do the same to us.
01:09:35
And right now, Chinese firms that are trying to develop AI
01:09:38
systems are doing so at least severe disadvantage because
01:09:41
they've got much less access to high quality hardware than does
01:09:45
Open AI or or Google. And that's I think the only way
01:09:50
to accelerate into the future as countries, including the US
01:09:53
become more restrictive in terms of controlling access to
01:09:56
different types of AI hardware. And so it's a good thing that
01:10:00
we've got most of it designed here and most of it manufactured
01:10:04
in friendly countries, because what we're going to see, what
01:10:07
we're already seeing, what we'll keep seeing is more
01:10:10
politicization, I think, of the hardware that makes AI possible.
01:10:13
Is the United? I mean the United States still
01:10:16
seems great. I guess that chip design, given
01:10:18
we have so many of the leading firms, could, do you think we
01:10:22
could ever produce a fab or like what?
01:10:24
What makes TSMC so uniquely capable at building a fab?
01:10:29
Like, we're talking culture, money, like, what's what's
01:10:34
driving that in your opinion? Yeah.
01:10:36
You know it's it's it's not it can't be a cultural thing
01:10:38
because Taiwan hasn't always been the center of the chip
01:10:41
industry. And if you go back four decades
01:10:43
ago, it was Japan that was the world's biggest producer of of
01:10:46
advanced semiconductors. So it's primarily I think the
01:10:50
result of TSMC business model which has let it scale and the
01:10:54
scale has let it Dr. efficiencies.
01:10:57
And so when TSMC was founded in 1987 at that time almost all
01:11:01
companies both designed and manufactured chips in house.
01:11:04
They did, both sides of the of the equation, And the founder of
01:11:08
TSMC, Morris Chang, realized that as it was getting more
01:11:11
complex to manufacture, companies would prefer to
01:11:14
outsource it. And so he decided that he wanted
01:11:16
to be sort of like Gutenberg was for books.
01:11:18
Gutenberg didn't write any books, he only printed them.
01:11:20
Morris Chang and TSMC, they didn't design any chips.
01:11:23
They only manufactured them. And what that meant is that he
01:11:25
could produce chips not just for one company, but for lots of
01:11:29
companies. For AMD, for Apple, for
01:11:32
Qualcomm, for NVIDIA. And so today, TSMC is the
01:11:34
world's largest chip maker. And precisely because it's the
01:11:37
largest, it's also the most advanced because it can hone, it
01:11:40
processes over every single silicon wafer that it
01:11:43
manufacturers. How can you can you make any
01:11:47
predictions about how much you think AI chips are going to
01:11:51
improve? Or like what do you see sort of
01:11:53
the next couple years looking like in terms of like do
01:11:57
availability problems resolve and and how much more do you
01:12:01
think these chips will improve over the next couple years?
01:12:06
Yeah, on the the availability front, NVIDIA and TSMC has said
01:12:10
they expect it to be a year or so until the shortages get
01:12:14
resolved, but I think they're going to get resolved.
01:12:17
Prices are such that there's a very strong incentive to solve
01:12:20
the availability problem and sell more.
01:12:23
Sometimes people think they want to keep a shortage to keep sort
01:12:27
of demand high. You you think they want to
01:12:29
produce as many chips as they can get out there to make the
01:12:31
money while it's there to be made or yeah I think that's
01:12:35
right because they're they're it looks like they're a
01:12:37
monopolistic producer. They've got 90% of the market
01:12:40
but if they can't supply chips there are competitors there is
01:12:43
Google with TPUS there is AMD. And so if there are long run
01:12:47
shortages that that are are in the supply chain, people will
01:12:50
turn to competitors. And so they've got a strong
01:12:52
incentive to get get supply increase.
01:12:54
And actually part of their strategy is again to put their
01:12:57
chips at the center of the AI world and make it the the gold
01:13:01
standard. And the more people that design,
01:13:04
the more people that train, more people that study using their
01:13:08
system using the Kuda ecosystem, the more likely NVIDIA is to be
01:13:11
at the center going forward. And then I guess sort of the
01:13:14
next chip they have is hard to know or have there been any
01:13:17
leaks about what what the future looks like in terms of these
01:13:21
artificial intelligence chips? I mean I think the the best
01:13:24
guide is is to look at the past and and what you find is that
01:13:28
that for over the past half decade they've released a series
01:13:33
of better and better chips that you know ballpark Moore's law
01:13:37
gives you a pretty good guide as to where things should be in a
01:13:40
couple of years time bigger, more powerful hopefully somewhat
01:13:44
more power efficient and and the other area focuses the the
01:13:49
interconnect speed between semiconductors.
01:13:52
So there's the GPUs themselves, there's all the networking
01:13:54
equipment that moves data between chips.
01:13:57
That's also an area of really intense focus right now.
01:14:01
And we're I think we're gonna see a lot of progress for the
01:14:03
coming years. Jensen I mean, did you interact
01:14:06
with him at all for this book or no, I didn't actually.
01:14:10
I I interviewed his his Co founder but not not.
01:14:12
Interesting. Yeah.
01:14:13
I mean, just like observe. Like it must be crazy to go from
01:14:18
sort of being seen as this, like, I don't know, wild sort of
01:14:21
futurist to all of it. Seeming to come true or the
01:14:25
world sort of agreeing with your futurism.
01:14:29
I mean he sort of he he wears like leather jackets and dresses
01:14:32
all black or like what? What do you make of him?
01:14:35
And yeah, how much has this been sort of a vindication for him?
01:14:40
Well, I think he he sort of sees himself like like Steve Jobs was
01:14:44
for the iPhone moment, he is for the AI moment.
01:14:47
And I think that's right. I I think he he has played that
01:14:50
role because he does with a lot of credit for investing in the
01:14:54
types of products that that have made AI possible at a time when
01:14:58
most people thought it was either crazy or possible but
01:15:02
only relevant for niche academic uses or something that would
01:15:07
never be financially viable. And look, he's built one of the
01:15:10
world's largest tech companies around it.
01:15:13
And already there are products being produced on a regular
01:15:16
basis using chips that his company has pioneered.
01:15:20
I believe I agree with you. You can make a fortune.
01:15:23
Silicon Valley can love you. But I feel like the world will
01:15:26
only love you if it's your device in their hands or it's
01:15:29
your brand. You know, it.
01:15:31
I, I, it just doesn't seem like they have a path for like
01:15:34
consumer love, right? I mean, do you see them trying
01:15:37
to take advantage of this and go, you know, offer other
01:15:42
products or, you know, they're so good at chips, they're
01:15:45
they're gonna, they'll win the technologist's heart and be
01:15:47
happy with their cash mile. Yeah.
01:15:51
They're so far from the computer in terms of where they sit in
01:15:54
the tech stack that like most consumers don't even know they
01:15:58
exist. Or gamers do is sort of the
01:16:02
funny gamers do. Right.
01:16:03
Yeah. How did how did you get into
01:16:07
this? Like what were you sorry, I
01:16:09
don't know as much of your background.
01:16:11
Yeah. Before you wrote this book, were
01:16:12
you a chip nerd or what? No, not at all.
01:16:15
Oh, really? OK what was your story to the
01:16:17
book? So I I'm a an economic historian
01:16:20
by training written a couple of books on different aspects of
01:16:22
economic history and and decided to write the book when I came to
01:16:26
realize how important ships were and was shocked myself that I'd
01:16:30
never paid any attention to this industry.
01:16:32
And so it was the combination of learning that almost all
01:16:36
advanced ships are made in Taiwan, which was shocking.
01:16:39
Learning that chips are so extraordinarily difficult to
01:16:42
manufacture and transistors in your iPhone, for example, are
01:16:46
small in the size of a coronavirus manufactured by the
01:16:49
billions. That was an extraordinary fact
01:16:52
that got me into it. And then and then finally
01:16:54
looking at the ramifications for AII, like most people have
01:16:57
thought about AI as a as a question of of of either
01:17:00
software or of data. Data is the new oil I'd read a
01:17:03
million times, but it turns out that's not exactly right.
01:17:06
Did you go dump all your money into NVIDIA or I should have?
01:17:11
Yeah. Did you see the stock?
01:17:13
I mean I I know. Well you're if you're an
01:17:15
economic historian like do you have a view on whether, OK, this
01:17:19
is a great company but it's overvalued.
01:17:21
I mean you know, yeah it's do you, I mean you shouldn't turn
01:17:24
to me for for stock market advice.
01:17:26
But I mean I think the, the key question about NVIDIA is, is
01:17:30
will it. I, I think it's clear it will
01:17:33
continue to play a dominant role in AI training.
01:17:36
Maybe it'll lose market share, maybe it'll win market share
01:17:38
with Google, but it'll play a central role.
01:17:40
But in the inference market will NVIDIA be as central there or
01:17:44
not? That's I think the key question
01:17:46
that we're gonna find the answer to over the next couple of
01:17:48
years. And do you think China is close
01:17:51
to catching up? You know, I I think China's is
01:17:53
not that far behind, but China has been not that far behind for
01:17:58
some time. And because the rate of advance
01:18:01
is so rapid in terms of design in the US and manufacturing in
01:18:04
Taiwan, it's very, very hard to catch up.
01:18:08
And so long as that companies like NVIDIA and TSMC keep racing
01:18:12
forward to catching up is just going to be extraordinarily
01:18:15
challenging for China and no matter how much money they pour
01:18:16
into it. And are Japan and South Korea
01:18:19
relevant to this conversation? Well, it's that's an interesting
01:18:22
question. Japan is relevant because if you
01:18:26
look inside of a chip making facility, there is a ton of
01:18:29
Japanese machine tools and there's also a wide variety of
01:18:34
ultra specialized chemicals made in Japan.
01:18:37
So a lot of the chemicals used in chip making because of the
01:18:40
precision required have to be purified to the 9999% level.
01:18:46
A lot of those are only made in Japan.
01:18:49
I wanted to go back and I know this is like a complicated
01:18:53
question, but like what are the tasks that are being thrown at
01:18:58
these chips, right? Or you talked about sort of
01:19:01
matrix computing or? Is there like 1 task over and
01:19:04
over again that these chips are being asked to do or it's a lot
01:19:07
of different things or how? How are the foundation models
01:19:11
sort of translating to the chips as best you can explain it?
01:19:16
I I guess the way I would explain it is you know if you're
01:19:18
trying to train a a a chat bot like ChatGPT, you train it by
01:19:24
taking all of the language data in a a vast data set like
01:19:27
Wikipedia. You essentially read Wikipedia
01:19:30
and identify patterns, and so and.
01:19:33
That's and that's like inference that you've been taught.
01:19:36
That's training. That's training.
01:19:37
So training is is identifying the pattern.
01:19:39
So you know the what's the what? What is the next word I'm going
01:19:48
to say? You know, you know what say, and
01:19:52
Chachi Piti knows that because it's identified thousands of
01:19:55
other instances where sentences like that ended with the word
01:19:59
say. And it did that because it read
01:20:02
lots and lots of books and articles and Reddit posts
01:20:06
online. And then when I you go back and
01:20:08
ask a question which is inference, it gives you a pretty
01:20:11
good answer because it has a pretty good sense of what the
01:20:14
most likely word is going to be. So you're saying they're they're
01:20:18
training the chips based on a a lot of data and then it makes an
01:20:23
inference about what it thinks the the sort of fill in the
01:20:26
blank of of the prompt? Basically, yeah, inference is
01:20:30
just an educated guess, a probabilistic guess.
01:20:33
And what's cool about Chachi PT is that it's not just fill in
01:20:37
one word, it's fill in dozens of.
01:20:39
Words. I know, it's insane.
01:20:41
One thing I'm still trying to get my head around is, you know,
01:20:44
like DeepMind, right? Like when they trained, you
01:20:50
know, computer systems, I guess to to compete on chess or Go or
01:20:55
Dota. Like my sense is that you know
01:20:58
they're using. Sort of this sort of scale data
01:21:01
approach, but they also have like specific ways they're
01:21:04
teaching it to think about specific problems.
01:21:08
Or like, do you, are those types of approaches to AI also using
01:21:13
these same chips, or are they still sort of invested in the
01:21:17
same arms race? Yeah, well, the chips just do
01:21:21
the math for you. And so they'll do what other
01:21:22
math? Whatever math you ask them to
01:21:24
do, Yeah. Exactly.
01:21:27
And they're not. And then the interaction between
01:21:31
like Nvidia's like programming language is super interesting
01:21:36
because they're not like the the foundation models are not being
01:21:38
written in that language, right? But they're they're sort of
01:21:40
bridging. That's right.
01:21:43
Yep. And it's it's just about, it's
01:21:46
fundamentally allows them to optimize what's happening on
01:21:48
these chips much more than maybe they would on like an AMD chip
01:21:52
even if it were and it's it's it's it's easier, it's more
01:21:55
familiar and therefore there's a big Moat around invidious
01:21:58
position. Yeah.
01:22:01
So I have you like become like were you bullish on like AI
01:22:06
before writing this book or I'm curious how much your view has
01:22:10
really changed on, yeah like I don't know like generalized
01:22:13
intelligence or like the imminence of self driving cars
01:22:16
or any of that. You know I I think that actually
01:22:19
we've I I'm I'm in the long run, I'm very bullish but I think in
01:22:23
in the the short run the last 12 months there's been all a whole
01:22:26
lot of excitement of the imminent arrival of AGI which
01:22:30
I'm I'm not so sure I see it happening in 2023.
01:22:34
I think if you look at the trends, the trends are all very
01:22:37
positive. AI system is getting better,
01:22:39
more capable, but I think there there there was due to ChatGPT a
01:22:44
popular sense that we're going to get some extraordinary
01:22:47
products in our hands right away.
01:22:48
And actually, it's going to be a couple of years before we see
01:22:51
AI, the current wave of AI products deployed in a big way
01:22:55
that meaningfully impacts economy or society.
01:22:57
Yeah, I think we're all watching to see if this hype wave is
01:23:00
about to end or not. I mean there's there's a world
01:23:03
where this is sort of peak NVIDIA, right?
01:23:06
Like we have this sort of mania around AI, everybody's running
01:23:10
towards it and then it turns out, you know, chat GPD doesn't
01:23:15
improve that much and self driving cars still have a lot of
01:23:18
edge cases. And you know, the chips don't
01:23:21
solve everything. Even if they do get better.
01:23:23
Like you're saying, like, I don't know.
01:23:24
How likely do you think that is? Or do you have a sense of, yeah,
01:23:27
whether better years are ahead or this is sort of a peak moment
01:23:31
for the company. Well it's it's been a very, very
01:23:34
good year for the company. So I wouldn't be surprised if
01:23:36
it's a it's it's at least a localized peak.
01:23:38
But I I think you know for for AI applications most people
01:23:43
think of what's the consumer application for AI, you know
01:23:46
what's what's the iPhone of AI. But it seems to me that for the
01:23:50
next couple of years actually most of the application and most
01:23:53
of the money in AI will be in enterprises because enterprises
01:23:56
that have vast data, they've got the desire to use it efficiently
01:23:59
to monetize it. And so that's where I think
01:24:02
we're actually already seeing a lot of the investment and a lot
01:24:04
of the early products are happening in pretty boring
01:24:07
places inside of enterprises. But that that's probably OK
01:24:10
that's that's what's going to drive I think the the, the
01:24:14
building of effective products in the long run.
01:24:17
Enterprises, you know, businesses just translate it for
01:24:19
like the regular person, you know, I used to work at
01:24:21
Bloomberg. They've actually come out, I
01:24:23
think with their own foundation model to show how they can train
01:24:26
financial data. And you you have a good example
01:24:29
of where you think corporations will put this approach and data
01:24:33
to use. If you imagine a company like
01:24:37
Walmart, they have to everyday decide what price to put all of
01:24:41
their products, and they decide pricing based on a whole variety
01:24:45
of different factors. What their competitors are
01:24:46
pricing? The availability of products.
01:24:48
Very complex process and that seems like a perfect use case
01:24:53
for trying to use more advanced AI algorithms to set better
01:24:57
prices more rapidly. Boring is it?
01:24:59
Kmart this blue light discounts. What's Walmart?
01:25:01
I forget. But yeah, who gets the discount?
01:25:03
You know where, right? A lot of products to keep track
01:25:07
of. Have you like?
01:25:10
It feels so absurd to ask this, but like, what the chip business
01:25:14
looks like in a world of AGI, generalized intelligence?
01:25:18
Or like, have you gained that out at all?
01:25:20
I mean, there are people who think it could happen, you know,
01:25:24
five years from now or less. Like, do you have a view on it
01:25:28
first of all and then second? Yeah, what?
01:25:30
What would it mean for the chip world?
01:25:34
Well I I guess my sense is that we're we're we're a long way
01:25:37
away. I'm I'm still waiting for
01:25:38
ChatGPT to accurately complete all my senses.
01:25:41
You know still at a 90% rate and the other 10% are pretty bad.
01:25:46
I think it's gonna require tremendous advances in
01:25:50
semiconductors to make possible the tremendous increase in
01:25:54
computing that more advanced AI systems will need.
01:25:56
I mean that that's been the trend.
01:25:57
The trend has been AI systems only advance.
01:26:00
We apply more computing to them, and so if we want systems to be
01:26:03
twice as good as they are today, we're gonna need something not
01:26:07
too far off from twice as much compute to make it possible.
01:26:10
Are there, you know there was this I I covered Righetti and
01:26:14
sort of the quantum computing world.
01:26:17
Do you have much optimism there? Do you spend time on them in in
01:26:21
the book? You know, I not not in the book.
01:26:25
I my sense in in, in, in speaking to people in the
01:26:29
quantum world and then looking also at how in history new
01:26:32
computing technologies have disseminated is that actually
01:26:36
even revolutionary technologies, they're implemented slowly.
01:26:40
And so suppose we get to the next couple of years, the first
01:26:43
practical use case of quantum computing.
01:26:45
It's going to be a years long process of beginning to
01:26:49
implement that in all sorts of different computing use cases.
01:26:51
And so I think we shouldn't expect to have, you know, a
01:26:54
quantum powered iPhone anytime soon.
01:26:56
It's it's insane to think like on the one hand we have.
01:27:01
We have public companies that people are sort of speculating
01:27:04
and betting on they yeah produce quantum computers but they
01:27:08
really do nothing practical. I I you can go with I you know I
01:27:13
swing back and forth. On the one hand it's like great
01:27:15
it's amazing that our system will invest in such bleeding
01:27:19
edge technology and give it a chance.
01:27:21
On the other hand for the shareholders, you know like
01:27:24
everybody could decide to give up on the effort with like
01:27:26
higher interest rates, it's, I don't know, it's a crazy
01:27:29
function of. The global economy, well, I I
01:27:33
think it's interesting that a lot of the companies that are
01:27:35
investing the heaviest are also big cloud computing operators.
01:27:39
And that's I think because they believe that quantum will be
01:27:42
actually most useful in a context where it's closely
01:27:46
interlinked with huge volumes of classical computing.
01:27:50
And so actually, we're going to need more advanced silicon chips
01:27:53
to make quantum practically applicable.
01:27:55
And do you know, is there any sense whether Quantum and the
01:27:58
GPU sync up, or where they fit in that story?
01:28:03
There There are many different paradigms for how you
01:28:06
specifically sync classical and quantum computing, but right now
01:28:11
nobody knows which of the many paradigms will win, if any of
01:28:15
them. To wrap it up, I mean, what do
01:28:18
you think the regular person should take from all this?
01:28:21
Like it feels very far. From their lives like they they
01:28:26
maybe don't even know what NVIDIA is.
01:28:28
Like they tried ChatGPT, their kids using it for homework.
01:28:32
It's not in their life yet. Like you know Silicon Valley
01:28:35
loves to get itself worked up about things that don't
01:28:38
sometimes translate. Don't like what what do you
01:28:40
think the lesson or the the key thing for like the regular
01:28:44
person right now is in terms of what's happening with these
01:28:47
chips. I think if you went back to
01:28:49
1965, which is when Gordon Moore first set out the the phrase
01:28:53
Moore's Law and you asked what's the impact of of of Moore's Law,
01:28:57
the average person, the the answer in the short run was
01:29:00
approximately nothing. But the answer in the long run
01:29:03
is that it totally transformed society, economy, technology,
01:29:06
everything, because we put computing and therefore
01:29:10
semiconductors into basically every product that we rely on.
01:29:13
And I think we should expect the same to be true for AI.
01:29:16
You know, what does it mean for me tomorrow?
01:29:19
Probably not much. And what is it going to mean in
01:29:21
in 10 years and in 20 years when every aspect of human life is
01:29:25
being impacted by it? Well, it'll be transformative in
01:29:28
all sorts of ways, most of which we probably can't even imagine
01:29:30
today. For the startup entrepreneur,
01:29:32
where do you think there's opportunity to build a business?
01:29:36
You know they, I mean maybe you think they should go out and
01:29:38
start the next NVIDIA like it's a heavy lift, but like where
01:29:42
where do you think really with the progress that we've made in
01:29:46
these chips that there's real business opportunities still?
01:29:51
Well, I think if you've got the idea of the next NVIDIA, you
01:29:54
should absolutely go do it. But I I think you're, you're
01:30:00
right that, you know, companies like NVIDIA, they they create
01:30:03
the infrastructure on which many different types of systems can
01:30:06
be built. And and if you go back to, you
01:30:09
know, the smartphone for example, smartphones were
01:30:11
themselves a platform owners could build lots of different
01:30:14
things. And I think we're still in the
01:30:16
Super early stages. We're basically in stage zero in
01:30:18
terms of figuring out what are the ways you can create products
01:30:21
out of generative AI systems. Right now, there are hardly any
01:30:24
companies that make money selling their AI systems.
01:30:27
And of course, the challenge for entrepreneurs is, you know,
01:30:30
people who built early iPhone apps and Facebook apps, you
01:30:33
know, they don't all pan out. Sometimes you come too early.
01:30:36
You can be right that a technology is transformational
01:30:40
and get the timing wrong. And I guess that's.
01:30:43
The sort of confluence of luck and insight in, in the business
01:30:47
world, it's a tough one to predict.
01:30:50
Chris, this was awesome. Thank you so much for coming on
01:30:53
the show. It's great.
01:30:54
Great to talk to you. Thanks for having me.
01:30:56
That's our episode on chips and big tech.
01:30:59
Thanks for listening. I'm Eric Newcomer, your host,
01:31:03
author of newcomer. Thanks so much to Max Child and
01:31:06
James Wilsterman. Co founder Zavali and my long
01:31:08
time friends shout out to Scott Brody, our producer Riley
01:31:12
Kinsella, my Chief of Staff Gabby Caliendo at Volley who's
01:31:15
helping organize the conference and playing a big role behind
01:31:19
the scenes. Thank you to young Chomsky for
01:31:21
the theme music. Please like, comment, subscribe
01:31:24
on YouTube, give me a review on Apple podcast and subscribe to
01:31:29
the sub stack newcomer.co. Go try out Volley on your Alexa
01:31:34
play song quiz maybe we're going to have.
01:31:36
Couple more episodes before Cerebral Valley.
01:31:39
Cerebral Valley is on November 15th.
01:31:42
We will publish our conversations to YouTube.
01:31:46
We will probably publish them all.
01:31:47
That's what we did last time. You can see the conversations
01:31:50
from Cerebral Valley One, which is on March 30th.
01:31:54
And then yeah, we'll play some of our favourites probably.
01:31:58
We did a sort of distilled version in our podcast feed, so
01:32:02
follow the podcast feeds for those.
01:32:04
And yeah, I'll be covering it in the newcomer newsletter.
01:32:07
So newcomer.co, Thanks so much.
