[00:00] (0.24s)
two things. One is how quickly starting
[00:05] (5.44s)
in March of 2023, so we're talking about
[00:09] (9.84s)
less than two and a half years, what
[00:12] (12.48s)
this team has accomplished
[00:15] (15.20s)
and how far ahead they are of everybody
[00:18] (18.16s)
else as demonstrated by this. But the
[00:20] (20.80s)
second is a fundamental architectural
[00:23] (23.68s)
decision that Elon made which I think we
[00:27] (27.44s)
didn't fully appreciate until now. And
[00:30] (30.32s)
it maps to an architectural decision he
[00:33] (33.68s)
made at Tesla as well. And for all we
[00:35] (35.60s)
know, we'll figure out that he made an
[00:37] (37.04s)
equivalent decision at SpaceX. And that
[00:40] (40.08s)
decision is really well encapsulated by
[00:43] (43.52s)
this essay, the bitter lesson by Rich
[00:46] (46.48s)
Sutton. And Nick, if you can just throw
[00:48] (48.40s)
this up here, but just to summarize what
[00:50] (50.48s)
this says, it basically says in a
[00:52] (52.24s)
nutshell
[00:53] (53.76s)
that you're always better off when
[00:56] (56.88s)
you're trying to solve an AI problem
[00:59] (59.12s)
taking a general learning approach that
[01:01] (61.36s)
can scale with computation because it
[01:03] (63.76s)
ultimately proves to be the most
[01:05] (65.36s)
effective. And the alternative would be
[01:08] (68.64s)
something that's much more human labored
[01:10] (70.80s)
and human involved that requires human
[01:13] (73.20s)
knowledge. And so the first method, what
[01:16] (76.80s)
it essentially allows you to do is view
[01:20] (80.24s)
any problem as an endless scalable
[01:24] (84.48s)
search or learning task. And as it's
[01:27] (87.92s)
turned out, whether it's chess or go or
[01:31] (91.20s)
speech recognition or computer vision,
[01:34] (94.08s)
whenever there was two competing
[01:35] (95.44s)
approaches, one that used general
[01:37] (97.12s)
computation and one that used human
[01:39] (99.44s)
knowledge, the general computation
[01:42] (102.16s)
problem always won. And so it creates
[01:46] (106.16s)
this bitter lesson for humans that want
[01:48] (108.24s)
to think that we are at the center of
[01:50] (110.72s)
all of this critical learning and all of
[01:52] (112.48s)
these leaps. And what these results show
[01:55] (115.92s)
is a general computational approach that
[01:58] (118.48s)
doesn't require as much human labeling
[02:00] (120.80s)
can actually get to the answer and
[02:02] (122.88s)
better answers faster. That has huge
[02:05] (125.20s)
implications because if you think about
[02:07] (127.12s)
all these other companies, what is Llama
[02:09] (129.28s)
been doing? They just spent 15 billion
[02:11] (131.28s)
to buy 49% of scale AI. That's exactly a
[02:14] (134.88s)
bet on human knowledge. What is Gemini
[02:17] (137.20s)
doing? What is OpenAI doing? What is
[02:18] (138.88s)
Anthropic doing? So all these things
[02:20] (140.88s)
come into question. I just think it's an
[02:22] (142.80s)
incredible moment in technology where we
[02:25] (145.28s)
see so many examples. Travis is another
[02:27] (147.36s)
one. What he's just talked about, you
[02:29] (149.36s)
know, the bitter lesson is you could
[02:32] (152.08s)
believe that, you know, food is this
[02:34] (154.56s)
immutable thing that's made meticulously
[02:36] (156.64s)
by hand by these individuals. Or you can
[02:38] (158.96s)
take this general purpose computer
[02:40] (160.32s)
approach, which is what he took, waited
[02:42] (162.24s)
for these cost curves to come into play,
[02:44] (164.24s)
and now you can scale food to every
[02:46] (166.08s)
human on Earth. I ju I just think it's a
[02:48] (168.32s)
it's so profoundly important.