[00:00] (0.16s)
It's really great to see all of you.
[00:02] (2.72s)
What I want to do today since this is
[00:05] (5.04s)
build as startup school is share with
[00:07] (7.68s)
you some lessons I've learned about
[00:09] (9.52s)
building startups at AI fund. AI funds a
[00:12] (12.08s)
venture studio and we build an average
[00:14] (14.00s)
of about one startup per month. And
[00:16] (16.56s)
because we co-founded startups, we're in
[00:18] (18.56s)
there writing code, talking about
[00:19] (19.84s)
customers, design on features, detering
[00:21] (21.52s)
pricing. And so we've done a lot of reps
[00:23] (23.60s)
of not just watching others build
[00:25] (25.20s)
startups, but actually being in the
[00:26] (26.80s)
weeds, building startups with
[00:28] (28.40s)
entrepreneurs. And what I want to do
[00:30] (30.48s)
today is um share with you some of the
[00:32] (32.64s)
lessons I've learned building startups,
[00:34] (34.48s)
especially around this changing AI
[00:36] (36.48s)
technology and what it enables. And
[00:38] (38.72s)
it'll be focused on the theme of speed.
[00:41] (41.92s)
So it turns out that for those of you
[00:44] (44.16s)
that want to build a startup, I think a
[00:45] (45.92s)
strong predictor for startup's odds of
[00:47] (47.92s)
success is execution speed. And I
[00:50] (50.48s)
actually have a lot of respect for the
[00:52] (52.16s)
entrepreneurs and executives that can
[00:54] (54.16s)
just do things really quickly and new AI
[00:58] (58.16s)
technology is enabling startups to go
[01:00] (60.56s)
much faster. So what I hope to do is
[01:03] (63.12s)
share with you some of those best
[01:04] (64.24s)
practices which are frankly changing
[01:06] (66.00s)
every two to three months still to let
[01:08] (68.08s)
you get that speed that hopefully lets
[01:09] (69.92s)
you have a higher odds of success.
[01:11] (71.76s)
Before diving to speed, you know, a lot
[01:13] (73.68s)
of people ask me, hey Andrew, where are
[01:15] (75.60s)
the opportunities for startup? So this
[01:17] (77.76s)
is what I think of as the AI stack where
[01:20] (80.88s)
at the lowest level are the
[01:22] (82.24s)
semiconductor companies then the clouds
[01:24] (84.56s)
are hyperscalers built on top of that. A
[01:27] (87.04s)
lot of the AI foundation model companies
[01:28] (88.72s)
built on top of that. And even though a
[01:31] (91.28s)
lot of the PR excitement and hype has
[01:33] (93.28s)
been on these uh technology layers, it
[01:36] (96.08s)
turns out that almost by definition, the
[01:38] (98.48s)
biggest opportunities have to be at the
[01:40] (100.48s)
application layer because we actually
[01:43] (103.28s)
need the applications to generate even
[01:45] (105.36s)
more revenue so that they can afford to
[01:47] (107.60s)
pay the foundation cloud and
[01:49] (109.92s)
semiconductor technology layers. So for
[01:52] (112.88s)
whatever reason media and social media
[01:54] (114.96s)
tends not to talk about the application
[01:57] (117.04s)
layer as much but for those of you think
[01:58] (118.80s)
you're building startups almost by
[02:00] (120.64s)
definition the biggest opportunities
[02:02] (122.56s)
have to be there although of course the
[02:04] (124.40s)
opportunities at all layers of the
[02:05] (125.92s)
stack. One of the things that's changed
[02:07] (127.84s)
a lot over the last year um and in terms
[02:10] (130.80s)
of AI tech trends I if you ask me what's
[02:13] (133.12s)
the most important tech trend in AI I
[02:15] (135.04s)
would say is the rise of agentic AI and
[02:18] (138.48s)
about a year and a half ago when I
[02:20] (140.16s)
started to go around and give talks to
[02:21] (141.84s)
try to convince people that AI agents
[02:24] (144.24s)
might be a thing I did not realize that
[02:27] (147.12s)
around last summer a bunch of marketers
[02:29] (149.76s)
would get a hold of this term and use it
[02:32] (152.16s)
as a sticker and slap it on everything
[02:33] (153.92s)
in site which made it almost lose some
[02:35] (155.92s)
of meaning but I want to share with you
[02:37] (157.76s)
from a technical perspective why I think
[02:40] (160.40s)
agentic AI is exciting and important and
[02:42] (162.96s)
also opens up a lot more startup
[02:45] (165.12s)
opportunities. So, it turns out that the
[02:47] (167.12s)
way a lot of us use LMS is to prompt it
[02:49] (169.92s)
to have it gener output. And the way we
[02:52] (172.32s)
have an LM output something is as if
[02:55] (175.12s)
you're going to a human or in this case
[02:57] (177.04s)
an AI and asking it to please type out
[02:59] (179.76s)
an essay for you by writing from the
[03:01] (181.84s)
first word to the last word all in one
[03:03] (183.92s)
go without ever using backspace. And
[03:06] (186.56s)
humans, we don't do our best writing,
[03:08] (188.48s)
being forced to type in this linear
[03:10] (190.24s)
order. And it turns out neither does AI.
[03:12] (192.64s)
But despite the difficulty of being
[03:14] (194.16s)
forced to write in this linear way, um
[03:16] (196.24s)
our LMS do surprisingly well. With
[03:18] (198.24s)
agentic workflows, we can go to AI
[03:20] (200.00s)
system and ask it to please first write
[03:22] (202.00s)
an essay outline, then do some webs
[03:24] (204.00s)
research if it needs to and fetch uh
[03:26] (206.24s)
some web pages to put in their own
[03:27] (207.84s)
context, then write a first draft, then
[03:30] (210.00s)
read the first draft and critique it and
[03:31] (211.68s)
revise it and so on. And so we end up
[03:33] (213.84s)
with this iterative workflow where your
[03:35] (215.76s)
model does some thinking and some
[03:37] (217.04s)
research does some revision goes back to
[03:39] (219.28s)
do more thinking and by going around
[03:41] (221.12s)
this loop many times. Uh it is slower
[03:43] (223.44s)
but it delivers a much better work
[03:45] (225.68s)
product. So for a lot of t lot of
[03:47] (227.84s)
projects AI fund has worked on
[03:49] (229.60s)
everything from um pulling out complex
[03:52] (232.72s)
compliance documents to uh medical
[03:55] (235.04s)
diagnosis to reasoning about complex
[03:57] (237.84s)
legal documents. We found that these
[03:59] (239.52s)
agentic workflows are really a huge
[04:01] (241.76s)
difference between it working versus not
[04:03] (243.92s)
working. But a lot of the work that
[04:06] (246.16s)
needs to be done, a lot of the valuable
[04:07] (247.52s)
businesses to be built still will be
[04:09] (249.28s)
taking workflows existing or new
[04:11] (251.04s)
workflows and figuring out how to
[04:12] (252.64s)
implement them into these size of
[04:14] (254.56s)
agentic workflows. So just to update the
[04:17] (257.84s)
picture for the AI stack, um what has
[04:20] (260.64s)
emerged over the last year is a new
[04:23] (263.12s)
agentic orchestration layer that helps
[04:26] (266.16s)
application builders orchestrate or
[04:28] (268.08s)
coordinate a lot of calls to the
[04:30] (270.16s)
technology layers underneath. And the
[04:32] (272.56s)
good news is uh the orchestration layer
[04:34] (274.72s)
has made it even easier to build
[04:36] (276.72s)
applications. But I think the basic
[04:38] (278.56s)
conclusion that the application layer
[04:40] (280.08s)
has to be the most valuable layer of the
[04:41] (281.68s)
stack still holds true with a bias or
[04:44] (284.88s)
focus on the application layer. Let me
[04:46] (286.96s)
now dive into some of the best practices
[04:48] (288.72s)
I've learned for how startups can move
[04:51] (291.04s)
faster.
[04:52] (292.96s)
It turns out that um at AI fun we only
[04:57] (297.12s)
focus on working on concrete ideas. So
[05:00] (300.40s)
to me a concrete idea a concrete product
[05:02] (302.88s)
idea is one that's specified in enough
[05:05] (305.20s)
detail that an engineer can go and build
[05:07] (307.60s)
it. So for example if you say let's use
[05:10] (310.56s)
AI to optimize healthcare assets you
[05:13] (313.20s)
know that's actually not a concrete
[05:14] (314.64s)
idea. It's too vague. If you tell me
[05:16] (316.40s)
it's a very software to use AI to
[05:17] (317.84s)
optimize healthcare assets different
[05:19] (319.60s)
engineers would do totally different
[05:21] (321.04s)
things and because it's not concrete you
[05:23] (323.28s)
can't build it quickly and you don't
[05:24] (324.80s)
have speed. In contrast, if you had a
[05:27] (327.20s)
concrete idea like let's write software
[05:28] (328.96s)
to let hospitals, let patients book MR
[05:31] (331.04s)
machine slots online to optimize usage.
[05:33] (333.04s)
I don't know if this is a good or a bad
[05:34] (334.64s)
concrete idea. Actually business is
[05:36] (336.00s)
already, you know, doing this. But it is
[05:38] (338.08s)
concrete and that means engineers can
[05:39] (339.68s)
build it quickly. If it's a good idea,
[05:41] (341.36s)
you find out it's not a good idea, you
[05:43] (343.28s)
will find out. But having concrete ideas
[05:45] (345.36s)
buys you speed. Or someone to say, let's
[05:47] (347.76s)
use AI for email personal productivity.
[05:49] (349.76s)
Too many interpretations of that. That's
[05:51] (351.52s)
not concrete. But if someone says could
[05:54] (354.00s)
you build an app Gmail integrate the
[05:55] (355.60s)
automation that use let's use the right
[05:57] (357.20s)
prom source right filter entire emails
[05:58] (358.88s)
that is concrete I could you know I
[06:00] (360.24s)
could go build that this afternoon. So
[06:02] (362.08s)
concretness buys you speed and the
[06:04] (364.80s)
deceptive thing for a lot of
[06:06] (366.24s)
entrepreneurs is the vague ideas tend to
[06:09] (369.04s)
get a lot of kudos. If you go and tell
[06:10] (370.88s)
all your friends we should use AI to
[06:12] (372.64s)
optimize the use of healthcare assets.
[06:14] (374.40s)
Everyone will say that's a great idea.
[06:16] (376.40s)
But it's actually not a great idea at
[06:17] (377.92s)
least in the sense of being something
[06:19] (379.52s)
you can build. Uh it turns out when
[06:21] (381.28s)
you're vague, you're almost always
[06:22] (382.56s)
right. Uh but when you're concrete, you
[06:25] (385.12s)
may be right or wrong. Either way is
[06:27] (387.12s)
fine. We can discover that much more
[06:28] (388.80s)
fast, which is what's important for SA.
[06:31] (391.20s)
In terms of executing on concrete ideas,
[06:33] (393.68s)
I I I find that at AI fun, I ask my team
[06:36] (396.16s)
to focus on concrete ideas because um a
[06:39] (399.04s)
concrete idea gives clear direction and
[06:41] (401.44s)
the team can run really fast to build it
[06:43] (403.52s)
and either validate it, prove it out or
[06:45] (405.36s)
falsify and conclude it doesn't work.
[06:47] (407.20s)
Either way is fine. So we can do that
[06:49] (409.28s)
quickly and it turns out that finding
[06:51] (411.76s)
good concrete ideas usually requires
[06:54] (414.16s)
someone could be you could be a subject
[06:55] (415.76s)
matter expert thinking about a problem
[06:57] (417.76s)
for a long time. Uh so for example
[06:59] (419.92s)
actually before before you know starting
[07:01] (421.84s)
Corsera um I spent years right thinking
[07:04] (424.96s)
about online education talking to users
[07:07] (427.28s)
holding my own intuitions about what
[07:09] (429.28s)
would make a good edtech platform and
[07:12] (432.08s)
then after that long process I think YC
[07:14] (434.40s)
sometimes calls it wondering the idea
[07:16] (436.00s)
maze but after thinking about it for a
[07:17] (437.60s)
long time you find that the guts of
[07:19] (439.60s)
people that have thought about this for
[07:21] (441.12s)
a long time can be very good about
[07:23] (443.12s)
rapidly making decisions as in after
[07:25] (445.36s)
you've thought about this talked to
[07:26] (446.48s)
customers and so on for a long time if
[07:28] (448.24s)
you ask this expert, should I build this
[07:29] (449.84s)
feature or that feature? You know, the
[07:32] (452.08s)
gut, which is an instantaneous decision,
[07:34] (454.40s)
uh, can be actually a surprisingly good
[07:36] (456.16s)
proxy. It can be surprisingly good
[07:38] (458.32s)
mechanism for making decisions. And I
[07:40] (460.72s)
know I work on AI, you might think I'll
[07:43] (463.44s)
say, oh, we need data. And of course, I
[07:45] (465.36s)
love data, but it turns out getting data
[07:48] (468.16s)
for a lot of startups is actually slow
[07:49] (469.92s)
mechanism for making decisions. And a
[07:52] (472.08s)
subject matter expert with a good gut is
[07:54] (474.00s)
often a much better mechanism for making
[07:56] (476.08s)
a speedy decision. And then one other
[07:58] (478.00s)
thing for many successful startups at
[08:00] (480.48s)
any moment in time you're pursuing one
[08:02] (482.80s)
very clear hypothesis they building out
[08:05] (485.28s)
and trying to sell China value of
[08:06] (486.56s)
hotspot. Um and a startup doesn't have
[08:09] (489.60s)
resources to hedge and try 10 things at
[08:11] (491.44s)
the same time. So pick one go for it and
[08:14] (494.16s)
if data tells you to lose faith in that
[08:16] (496.56s)
idea that's actually totally fine. Just
[08:18] (498.56s)
pivot on a dime to pursue a totally
[08:20] (500.88s)
different concrete idea. So that's what
[08:22] (502.48s)
often feels like an AI fund. We're
[08:24] (504.24s)
pursuing one thing doggedly with
[08:26] (506.40s)
determination until the world tells us
[08:28] (508.64s)
we were wrong then change and pursue a
[08:30] (510.72s)
totally different thing with equal
[08:32] (512.08s)
determination and equal doggedness. And
[08:34] (514.88s)
one other pattern I've seen, if every
[08:37] (517.36s)
piece of new data causes you to pivot,
[08:39] (519.52s)
it probably means you're starting off
[08:41] (521.36s)
from two weaker a base of knowledge,
[08:42] (522.88s)
right? If every time you talk to a
[08:44] (524.00s)
customer, you totally change your mind,
[08:45] (525.68s)
probably means you don't know enough
[08:47] (527.20s)
about that sector yet to have a really
[08:48] (528.64s)
high quality concrete idea and uh
[08:51] (531.36s)
finding someone that's thought about a
[08:52] (532.72s)
subject for longer may get you on to
[08:54] (534.64s)
better path in order to go faster. The
[08:57] (537.20s)
other thing I often think about is the
[08:58] (538.80s)
built feedback loop which is rapidly
[09:00] (540.80s)
changing when it comes to how we build
[09:02] (542.88s)
with AI coding assistance. So when
[09:05] (545.76s)
you're building a lot of applications,
[09:07] (547.36s)
one of the biggest risks is custom
[09:08] (548.80s)
acceptance, right? A lot of startups
[09:10] (550.40s)
struggle not because we can't build
[09:12] (552.24s)
whatever we want to build, but because
[09:13] (553.44s)
we build something and it turns out
[09:15] (555.12s)
nobody cares. And so for a lot of the
[09:17] (557.84s)
way uh you know I build startups
[09:19] (559.76s)
especially applications less so deep
[09:21] (561.36s)
tech less so technology startups but
[09:23] (563.28s)
definitely application startups is we
[09:25] (565.28s)
often build software so this is an
[09:27] (567.12s)
engineuring toss and then we will get
[09:29] (569.68s)
feedback from users and this is a
[09:31] (571.28s)
product management toss and then we'll
[09:33] (573.44s)
go back you know then based on the user
[09:35] (575.20s)
feedback we'll tweak our views on what
[09:36] (576.88s)
to build go back to write more software
[09:39] (579.04s)
and we go around this loop many many
[09:40] (580.64s)
times iterate toward product market fit
[09:43] (583.20s)
and it turns out that with AI coding
[09:46] (586.80s)
assistance which Andre talked about as
[09:48] (588.56s)
well um rapid engineering is becoming
[09:52] (592.08s)
possible in a way that just was not
[09:53] (593.52s)
posing much more feasible. So the speed
[09:56] (596.72s)
of engineing is going up rapidly and the
[09:59] (599.12s)
cost of engineing is also going down
[10:01] (601.52s)
rapidly. This changes the mechanisms by
[10:04] (604.16s)
which we drive startups around this
[10:06] (606.08s)
loop. When I think about the software
[10:08] (608.08s)
that I do, I maybe put into two major
[10:09] (609.84s)
buckets. Sometimes I'll build quick and
[10:11] (611.68s)
dirty prototypes to test an idea. You
[10:13] (613.44s)
say build a new customer service
[10:14] (614.96s)
chatbot. Let's build AI to process legal
[10:16] (616.80s)
documents whatever build a quick and
[10:18] (618.32s)
dirty prototype to see if we think it
[10:20] (620.00s)
works. The other type of software where
[10:21] (621.84s)
I do is write maintain production
[10:23] (623.92s)
software maintain legacy software but
[10:26] (626.24s)
these massive production ready code
[10:28] (628.08s)
bases depending on which analysts report
[10:30] (630.56s)
you trust. It's been hard to find very
[10:32] (632.24s)
rigorous data on this. You know when
[10:34] (634.32s)
writing production quality code maybe
[10:36] (636.32s)
we're 30 to 50% faster with AI systems.
[10:38] (638.80s)
Hard to find a rigorous number. maybe
[10:40] (640.56s)
he's plausible to be but in terms of
[10:42] (642.48s)
building quick and dirty prototypes
[10:44] (644.40s)
we're not 50% faster I think we're
[10:46] (646.88s)
easily 10 times faster maybe much more
[10:48] (648.88s)
than 10 times faster and there are a few
[10:51] (651.36s)
reasons for this uh when you're building
[10:53] (653.36s)
standalone prototypes there's less
[10:55] (655.36s)
integration with legacy software
[10:57] (657.44s)
infrastructure legacy data needed um
[11:00] (660.40s)
also the requirements reliability even
[11:03] (663.36s)
scalability even security are much lower
[11:06] (666.32s)
and I know I'm not supposed to tell
[11:07] (667.76s)
people to write insecure code right
[11:09] (669.68s)
feels like the wrong thing to say, but I
[11:11] (671.36s)
routinely go to my team and say, "Go
[11:12] (672.80s)
ahead, write insecure code." Because if
[11:14] (674.80s)
this software is only going to run on
[11:16] (676.48s)
your laptop and you don't plan to
[11:18] (678.72s)
maliciously hack your own laptop, it's
[11:20] (680.88s)
fine to have insecure code, right? But
[11:22] (682.88s)
of course, after it seems to be working,
[11:24] (684.48s)
please do make it secure before you ship
[11:26] (686.08s)
it to someone else. And you know, like a
[11:27] (687.84s)
leaking PI, leaking sense data that that
[11:30] (690.24s)
is, you know, very damaging. So before
[11:32] (692.16s)
you ship it, make it secure and
[11:33] (693.36s)
scalable, but they're just testing it.
[11:34] (694.96s)
It's fine. And so I find increasingly
[11:37] (697.12s)
startups will systematically pursue
[11:39] (699.44s)
innovations by building 20 prototypes to
[11:41] (701.92s)
see what works, right? Uh because I I I
[11:45] (705.12s)
know that there's some angst in AI. A
[11:46] (706.64s)
lot of proof of concepts don't make to
[11:48] (708.08s)
production. But I think by driving the
[11:50] (710.24s)
cost of a proof of concept low enough,
[11:51] (711.84s)
it's actually fine if lots of proof of
[11:53] (713.36s)
concepts don't see the light of day. And
[11:56] (716.00s)
I know that um the mantra move fast and
[11:58] (718.72s)
break things got a bad rep because you
[12:00] (720.80s)
know it broke things. And some teams
[12:03] (723.36s)
took away from this that you should not
[12:05] (725.76s)
move fast, but I think that's a mistake.
[12:08] (728.40s)
I tend to tell my teams to move fast and
[12:10] (730.48s)
be responsible. And I think they
[12:11] (731.68s)
actually lots of ways to move really
[12:13] (733.44s)
quickly while still being responsible.
[12:16] (736.00s)
And in terms of the AI assistance uh
[12:19] (739.20s)
coding landscape, I think was it three
[12:21] (741.52s)
four years ago code autocomplete right
[12:23] (743.76s)
popularized by GitHub copilot and then
[12:26] (746.24s)
there was a cursor windserve generation
[12:28] (748.40s)
of AI enabled ids which great use winds
[12:31] (751.36s)
and cursor quite a lot um and then
[12:33] (753.84s)
starting I don't know six seven months
[12:36] (756.48s)
ago uh there started to be this new
[12:38] (758.64s)
generation of highly agentic coding
[12:40] (760.64s)
assistants uh including that she's using
[12:43] (763.20s)
o3 a lot for coding um cloud code is
[12:45] (765.60s)
fant Fantastic. Since quad 4 release,
[12:47] (767.36s)
it's become and and ask me again in a
[12:49] (769.28s)
few months, I may use something
[12:50] (770.24s)
different. But the tools are evolving
[12:51] (771.84s)
really rapidly, but I think uh cloud
[12:54] (774.32s)
codeex this is a new generation of
[12:56] (776.16s)
highly agentic coding assistance that is
[12:58] (778.56s)
making developer productivity keep on
[13:00] (780.64s)
growing. And the interesting thing is if
[13:02] (782.16s)
you're even half a generation or one
[13:03] (783.76s)
generation behind actually makes a big
[13:05] (785.84s)
difference compared to if you're on top
[13:07] (787.28s)
of the latest tools and I find my team
[13:09] (789.60s)
is taking really different approaches to
[13:11] (791.36s)
software engineing now compared to even
[13:12] (792.96s)
three or six months ago. One surprising
[13:15] (795.28s)
thing is we we're used to thinking of
[13:17] (797.44s)
code as this really valuable artifact
[13:19] (799.28s)
because it's so hard to create but
[13:21] (801.44s)
because the cost of software engine is
[13:22] (802.96s)
going down code is much less of a
[13:24] (804.80s)
valuable artifact as it used to. So I'm
[13:26] (806.48s)
on teams where you know we've completely
[13:28] (808.48s)
rebuilt a codebase three times the last
[13:30] (810.32s)
month right because it's not that hard
[13:32] (812.08s)
anymore to just completely rebuild a
[13:33] (813.76s)
codebase pick a new data schema is fine
[13:36] (816.08s)
because the cost of doing that has
[13:38] (818.00s)
plummeted some of you may have heard of
[13:39] (819.76s)
Jeff Bezos's terminology of a two-way
[13:42] (822.00s)
door versus a one-way door. A two-way
[13:44] (824.08s)
door is decision that you can make. If
[13:45] (825.60s)
you change your mind, come back out, you
[13:47] (827.36s)
know, reverse it relatively cheaply.
[13:49] (829.20s)
Whereas a one-way door is you make a
[13:51] (831.12s)
decision and you change your mind is
[13:52] (832.56s)
very costly or very difficult to
[13:54] (834.08s)
reverse. So choosing the software
[13:56] (836.08s)
architecture of your tech stack used to
[13:58] (838.00s)
be a one-way door. Once you built on top
[13:59] (839.60s)
of a certain tech stack, you set a
[14:01] (841.84s)
database schema, really hard to change
[14:03] (843.60s)
it. So that used to be a one-way door. I
[14:05] (845.76s)
don't want to say it's totally a two-way
[14:07] (847.36s)
door, but I find that um my team will
[14:10] (850.24s)
more often build on a certain tech stack
[14:12] (852.40s)
a week later, change your mind, let's
[14:14] (854.40s)
throw the code base away and redo it
[14:16] (856.00s)
from scratch on a new tech stack. I I
[14:18] (858.00s)
don't want to overhype it. We don't do
[14:19] (859.44s)
that all the time. There are still costs
[14:21] (861.44s)
to redoing that. But I find my team is
[14:23] (863.36s)
often rethinking what is a one-way door
[14:25] (865.44s)
and what's now a two-way door because
[14:27] (867.20s)
the cost of software engineering is so
[14:28] (868.88s)
much lower now. And maybe going a little
[14:30] (870.80s)
bit beyond software engineering, I I I
[14:32] (872.80s)
feel like this actually a good time to
[14:33] (873.92s)
empower everyone to build of AI. Uh over
[14:36] (876.96s)
the last year, a bunch of people advised
[14:39] (879.68s)
others not to learn to code on the
[14:41] (881.12s)
grounds of AI were automated. I think
[14:43] (883.36s)
we'll look back on this as some of the
[14:45] (885.60s)
worst career advice ever given because
[14:49] (889.52s)
as better tools make software engineing
[14:52] (892.24s)
easier, more people should do it, not
[14:54] (894.16s)
fewer. So when many decades ago the
[14:56] (896.88s)
world moved from punch calls to keyboard
[14:58] (898.40s)
and terminal that made coding easier.
[15:00] (900.64s)
When we moved from assemblies high level
[15:02] (902.16s)
languages like cobalt um there actually
[15:04] (904.40s)
people arguing back then that now we
[15:06] (906.96s)
have cobalt we don't need programmers
[15:09] (909.12s)
anymore like people actually wrote
[15:10] (910.32s)
papers to that effect but of course that
[15:12] (912.32s)
was wrong and programming languages made
[15:14] (914.24s)
it easier to code and more people learn
[15:15] (915.84s)
to code text IDs ID is the AI coding
[15:19] (919.44s)
assistant um and as coding becomes
[15:21] (921.76s)
easier more people should learn to code.
[15:24] (924.32s)
I have a controversial opinion which is
[15:26] (926.96s)
uh I think actually it's time for
[15:28] (928.24s)
everyone of every job role to learn to
[15:30] (930.32s)
code and in fact on my team you know my
[15:32] (932.56s)
CFO my head of talent my recruiters my
[15:35] (935.44s)
front desk uh uh person all of them know
[15:38] (938.64s)
how to code and I actually see all of
[15:40] (940.08s)
them performing better at all of their
[15:42] (942.16s)
job functions because they can code and
[15:44] (944.32s)
I think um I'm probably a little bit
[15:45] (945.76s)
ahead of the curve probably most
[15:46] (946.88s)
businesses are not there yet but in the
[15:48] (948.72s)
future I think we empower everyone to
[15:50] (950.56s)
code a lot of people can be more
[15:52] (952.48s)
productive I want to share with you One
[15:54] (954.08s)
lesson I learned as well on on why we
[15:56] (956.56s)
should have people learn to do this
[15:58] (958.48s)
which is um when I was teaching
[16:00] (960.32s)
generative VI for everyone on Corsera,
[16:02] (962.72s)
we needed to generate background art
[16:04] (964.64s)
like this uh using midjourney and you
[16:07] (967.44s)
know one of my team members uh new art
[16:10] (970.32s)
history and so he could prompt
[16:12] (972.40s)
midjourney with the genre the palette
[16:15] (975.20s)
the artistic inspiration had a very good
[16:17] (977.04s)
control over the images he generated. So
[16:19] (979.20s)
we end up using all of Tommy's generated
[16:21] (981.44s)
images. Whereas in contrast, I don't
[16:24] (984.40s)
know art history. And so when I prompt,
[16:27] (987.36s)
you know, image generation, I could
[16:29] (989.44s)
write, please make pretty pictures of
[16:31] (991.92s)
robots for me, right? And and I could
[16:34] (994.24s)
never have the control that my
[16:36] (996.24s)
collaborators could. And so I couldn't
[16:38] (998.08s)
generate as good images as he could. And
[16:40] (1000.56s)
I think with computers, one of the most
[16:43] (1003.20s)
important skills of the future is the
[16:44] (1004.88s)
ability to tell a computer exactly what
[16:46] (1006.88s)
you want. So they'll do it for you. And
[16:48] (1008.96s)
will be people that have that deeper
[16:50] (1010.48s)
understanding of computers that will be
[16:52] (1012.40s)
able to command a computer to get the
[16:54] (1014.32s)
outcome you want. And learning to code,
[16:56] (1016.40s)
not not that you need to write the code
[16:57] (1017.84s)
yourself. Steer AI to code for you seems
[17:00] (1020.56s)
like it will remain the best way to do
[17:01] (1021.92s)
that for a long time. with software
[17:04] (1024.08s)
engineering becoming much faster. The
[17:07] (1027.04s)
other interesting dynamic I'm seeing is
[17:08] (1028.72s)
that the product management work getting
[17:10] (1030.80s)
user feedback deciding what features to
[17:12] (1032.64s)
build that is increasingly the
[17:14] (1034.88s)
bottleneck and so I'm seeing very
[17:17] (1037.04s)
interesting dynamics in multiple teams
[17:18] (1038.72s)
over the last year a lot more of my
[17:20] (1040.16s)
teams have started to complain that
[17:21] (1041.36s)
their bottlenecks on product engineering
[17:22] (1042.88s)
and design because the engineers have
[17:24] (1044.56s)
gotten so much faster some interesting
[17:26] (1046.96s)
trends I'm seeing three four five years
[17:29] (1049.04s)
ago Silicon Valley used to have these
[17:31] (1051.44s)
slightly suspicious rules of thumb but
[17:33] (1053.04s)
nonetheless rules of thumb will have 100
[17:35] (1055.76s)
p.m. to four engineers or 1 PM to seven
[17:38] (1058.08s)
engineers was this like PM product
[17:40] (1060.08s)
manager to engineering ratio right which
[17:42] (1062.08s)
should take with a grain of salt but it
[17:43] (1063.60s)
was typical of a 1 PM to six seven
[17:45] (1065.92s)
engineers and with engineers becoming
[17:48] (1068.88s)
much faster I don't see product
[17:50] (1070.56s)
management work designing what to build
[17:52] (1072.48s)
becoming faster at the same speed
[17:54] (1074.72s)
engineers I'm seeing this ratio shift so
[17:57] (1077.84s)
literally yesterday one of my teams came
[17:59] (1079.76s)
to me and for the first time when we're
[18:01] (1081.76s)
planning headcom for a project this team
[18:03] (1083.76s)
proposed to me not at 1:00 PM to four
[18:06] (1086.32s)
engineers but to have 1 PM to 0.5
[18:09] (1089.12s)
engineers. So the team actually proposed
[18:10] (1090.48s)
to me I still know no this is a good
[18:12] (1092.16s)
idea for the first time in my life that
[18:13] (1093.84s)
I saw you know managers proposed to me
[18:16] (1096.08s)
having twice as many PMs as engineers
[18:18] (1098.40s)
was a very interesting dynamic. I I
[18:20] (1100.16s)
still don't know if this proposal I
[18:21] (1101.60s)
heard yes is a good idea but I think
[18:22] (1102.96s)
it's a sign of where the world is going
[18:25] (1105.44s)
and I find as PMs that can code or
[18:27] (1107.60s)
engineers with some product instincts
[18:29] (1109.04s)
often end up doing better. The other
[18:31] (1111.04s)
thing that I found important for startup
[18:32] (1112.64s)
found for startup leaders is because
[18:34] (1114.80s)
engineing is going so fast. If you have
[18:36] (1116.80s)
good tactics for getting rapid feedback
[18:38] (1118.64s)
to shape your perspective what to build
[18:40] (1120.40s)
faster that helps you get faster as
[18:43] (1123.28s)
well. So um I'm going to go through a
[18:45] (1125.68s)
portfolio of tactics for you know
[18:47] (1127.28s)
getting product feedback to keep shaping
[18:49] (1129.52s)
what you will decide to build. And we're
[18:51] (1131.84s)
going to go through a list of the faster
[18:54] (1134.00s)
maybe less accurate the slower more
[18:55] (1135.84s)
accurate tactics. So the fastest tactic
[18:59] (1139.12s)
for getting feedback is look at the
[19:00] (1140.64s)
product yourself and just go by your
[19:02] (1142.16s)
gut. And if you're a subject matter
[19:04] (1144.00s)
expert, this is actually surprisingly
[19:06] (1146.48s)
good you if you know what you're doing.
[19:08] (1148.40s)
A little bit slower is go ask three
[19:10] (1150.80s)
friends or teammates to get feedback to
[19:13] (1153.52s)
play with your product and get feedback.
[19:15] (1155.28s)
Um little bit slower is ask three to 10
[19:17] (1157.52s)
strangers you know for feedback. Um, it
[19:20] (1160.00s)
turns out for when I built products, one
[19:22] (1162.24s)
of the most important skills I think I
[19:23] (1163.92s)
learned was how to sit in the coffee
[19:25] (1165.68s)
shop, how to sit in a when it's
[19:27] (1167.52s)
traveling, when I travel, I often sit in
[19:29] (1169.04s)
the hotel lobby. It turns out learn to
[19:31] (1171.12s)
spot places of high foot traffic and
[19:33] (1173.36s)
very respectfully, you know, grab
[19:35] (1175.36s)
strangers and ask them for feedback on
[19:37] (1177.28s)
whatever I'm building. This used to be
[19:38] (1178.80s)
easier when I was less known. When when
[19:40] (1180.88s)
people recognize you, it's a little bit
[19:42] (1182.48s)
more awkward. I found that um I've
[19:44] (1184.80s)
actually sat with teams the hotel lobby
[19:46] (1186.56s)
very high foot for traffic and you know
[19:48] (1188.40s)
very respectfully ask strangers hey
[19:50] (1190.24s)
we're building this thing do you mind
[19:51] (1191.44s)
taking a look oh and I actually learned
[19:53] (1193.36s)
in a coffee shop there a lot of people
[19:55] (1195.84s)
working a lot of people don't want to be
[19:57] (1197.84s)
working so we give them excuse to be
[19:59] (1199.36s)
distracted they're very happy to do that
[20:01] (1201.20s)
too but I've actually kind of made tons
[20:03] (1203.68s)
of product decisions in a hotel lobby or
[20:05] (1205.52s)
a coffee shop with collaborators just
[20:07] (1207.36s)
just just like that send prototypes to
[20:09] (1209.36s)
100 testers you if you have access to
[20:11] (1211.68s)
logic group of users and prototype to
[20:13] (1213.76s)
more users and these are these get to be
[20:16] (1216.08s)
slow and slower tactics and I know
[20:18] (1218.00s)
Silicon Valley you know we like to talk
[20:19] (1219.76s)
about AB testing of course I do a ton of
[20:21] (1221.68s)
AB testing but contrary to what many
[20:24] (1224.32s)
people think AB testing is now one of
[20:26] (1226.08s)
the slowest tactics in my menu because
[20:28] (1228.48s)
it's just slow to ship it yeah it depend
[20:30] (1230.96s)
on how many users you have right so and
[20:32] (1232.96s)
then uh the other thing is um as you use
[20:35] (1235.36s)
anything but the first tactic some teams
[20:37] (1237.92s)
will look at the data they make a
[20:39] (1239.76s)
decision but the missing piece is When I
[20:42] (1242.72s)
AB test something, um, I don't just use
[20:46] (1246.24s)
result of AB test to pick product A or
[20:48] (1248.08s)
product B. My team will often sit down
[20:50] (1250.08s)
and look carefully at the data to hone
[20:52] (1252.00s)
our instincts to speed up to improve the
[20:54] (1254.32s)
rate. I wish we're able to use the first
[20:56] (1256.00s)
tactic to make high quality decision.
[20:58] (1258.40s)
Often sit down and think, gee, I
[21:00] (1260.40s)
thought, you know, this product name
[21:01] (1261.92s)
will work better than her product name.
[21:03] (1263.52s)
Clearly, my mental model the users
[21:05] (1265.28s)
wrong. to really sit down and think to
[21:07] (1267.76s)
update our mental model using all of
[21:10] (1270.00s)
that data to improve the quality of our
[21:13] (1273.12s)
guts on how to make product decisions
[21:15] (1275.12s)
faster. That turns out to be really
[21:16] (1276.80s)
important. All right, so talked about um
[21:20] (1280.24s)
concrete ideas, speed up engineering,
[21:22] (1282.32s)
speed up product feedback. This is one
[21:23] (1283.76s)
last thing I want to touch on which is
[21:25] (1285.36s)
I've seen that understanding AI actually
[21:27] (1287.36s)
makes you go faster. Um and and here's
[21:29] (1289.60s)
why. As a AI person, maybe I'm biased to
[21:31] (1291.44s)
be pro AI, but I want to share you why.
[21:33] (1293.52s)
So it turns out that when it comes to
[21:34] (1294.88s)
mature technology like mobile, you know,
[21:37] (1297.20s)
many people have had smartphones for a
[21:39] (1299.20s)
long time. We kind of know what a mobile
[21:40] (1300.72s)
app can do, right? So many people
[21:42] (1302.48s)
including nontechnical people have good
[21:44] (1304.16s)
instincts about what a mobile app can
[21:45] (1305.68s)
do. If you look at mature job roles like
[21:47] (1307.52s)
sales, marketing, HR, legal, they're all
[21:49] (1309.20s)
really important and all really
[21:50] (1310.40s)
difficult. But you know, there are
[21:52] (1312.64s)
enough marketers that that have done
[21:55] (1315.04s)
marketing for long enough and the
[21:56] (1316.72s)
marketing tactics haven't changed that
[21:58] (1318.64s)
much in the last year. So there are a
[22:00] (1320.72s)
lot of people that are really good in
[22:02] (1322.40s)
marketing and it's really important
[22:03] (1323.68s)
really hard but that knowledge is
[22:05] (1325.36s)
relatively diffused because you know the
[22:07] (1327.44s)
knowledge of how to do HR like it hasn't
[22:09] (1329.52s)
changed dramatically you know in the
[22:11] (1331.36s)
last six months but AI is emerging
[22:13] (1333.68s)
technology and so the knowledge of how
[22:16] (1336.16s)
to do AI really well is not widespread
[22:18] (1338.80s)
and so teams that actually get it that
[22:20] (1340.72s)
understand AI do have a advantage over
[22:23] (1343.68s)
teams that don't whereas if you need if
[22:25] (1345.20s)
you have an HR problem you can find
[22:26] (1346.80s)
someone you know that knows how to do it
[22:28] (1348.40s)
well probably but If an AI problem,
[22:30] (1350.56s)
knowing how to actually do that could
[22:32] (1352.56s)
put you ahead of other companies. So
[22:34] (1354.48s)
things like what accuracy can you get
[22:36] (1356.16s)
for a customer service chatbot? You
[22:37] (1357.60s)
know, should you prom fine tune a
[22:39] (1359.36s)
workflow? Um how do you get a voice out
[22:41] (1361.12s)
to low latency? There a lot of these
[22:42] (1362.72s)
decisions that if you make the right
[22:45] (1365.12s)
technical decision, you can like solve
[22:47] (1367.04s)
the problem in a couple days. They make
[22:48] (1368.96s)
the wrong technical decision, you could
[22:50] (1370.72s)
chase a blind alley for three months,
[22:52] (1372.64s)
right? And and one one thing I've been
[22:54] (1374.16s)
surprised by, it turns out if you have,
[22:56] (1376.00s)
you know, two possible architecture
[22:58] (1378.08s)
decisions, it's one bit of information.
[23:00] (1380.32s)
It feels like if you don't know the
[23:02] (1382.00s)
right answer, at most you're twice as
[23:03] (1383.76s)
slow, right? One bit, you know, try
[23:05] (1385.52s)
both. It feels like one bit of
[23:06] (1386.96s)
information can at most buy you a 2x
[23:08] (1388.96s)
speed up. And I think in some
[23:11] (1391.04s)
theoretical sense that is true. But what
[23:13] (1393.44s)
I see in practice, if you flip the wrong
[23:15] (1395.28s)
bit, you're not twice as slow. You spend
[23:17] (1397.60s)
like 10 times longer chasing a blind
[23:19] (1399.52s)
alley. which is why I think going in to
[23:21] (1401.68s)
have this right technical judgment, it
[23:23] (1403.76s)
really makes startups go so much faster.
[23:26] (1406.24s)
The other reason why I find staying on
[23:29] (1409.36s)
top of AI really helpful for startups is
[23:32] (1412.56s)
um over the last two years we have just
[23:36] (1416.64s)
had a ton of wonderful genai tools or
[23:39] (1419.68s)
genai building blocks right partial list
[23:42] (1422.64s)
but prompting workflows evals guardrails
[23:45] (1425.04s)
rack voice act async programming lots of
[23:47] (1427.04s)
ETL embeddings fine-tuning graph DB how
[23:49] (1429.76s)
to integ
[23:52] (1432.32s)
models there's a long and wonderful list
[23:54] (1434.32s)
of building blocks that can quickly
[23:56] (1436.32s)
combine to build software that no one on
[23:58] (1438.64s)
the planet could have built, you know,
[24:00] (1440.24s)
even a year ago. And this creates a lot
[24:02] (1442.32s)
of new opportunities for starters to
[24:04] (1444.08s)
build new things. So when I learned
[24:06] (1446.56s)
about these building blocks, this is
[24:07] (1447.76s)
actually a picture that I have in mind.
[24:09] (1449.52s)
If you own one building block, like you
[24:12] (1452.32s)
have a basic white building block, yeah,
[24:14] (1454.64s)
you can build some cool stuff. Maybe you
[24:16] (1456.24s)
know how to prompt. So you have one
[24:17] (1457.52s)
building block, you build some amazing
[24:19] (1459.04s)
stuff. But if you get a second building
[24:21] (1461.36s)
block like you also know how to build
[24:23] (1463.20s)
chat bots. So you have a white Lego
[24:25] (1465.28s)
brick and a black Lego brick, you can
[24:27] (1467.20s)
build something more interesting. Um if
[24:29] (1469.52s)
you acquire a blue building brick as
[24:31] (1471.60s)
well, you can build something even more
[24:33] (1473.20s)
interesting. Get few red building bras,
[24:36] (1476.08s)
maybe a little yellow one, more
[24:38] (1478.40s)
interesting, get more building bras, get
[24:40] (1480.88s)
more building bras, and very rapidly the
[24:43] (1483.36s)
number of things you comb combine them
[24:45] (1485.20s)
to into grows kind of combinatorily or
[24:48] (1488.16s)
grows exponentially. And so knowing all
[24:50] (1490.08s)
these wonderful building blocks lets you
[24:52] (1492.16s)
combine them in much richer combination.
[24:54] (1494.88s)
One of the things that deep learn does
[24:56] (1496.24s)
so I actually take a lot of deep learn
[24:57] (1497.92s)
courses myself you know to because work
[24:59] (1499.76s)
with great we work with I think like
[25:01] (1501.20s)
pretty much all the leading AI companies
[25:03] (1503.20s)
in the world and cert and and um and uh
[25:06] (1506.24s)
try to hand out building blocks. Um but
[25:09] (1509.12s)
when I look at the deep learning course
[25:11] (1511.12s)
catalog this is actually what I see. And
[25:13] (1513.52s)
whenever I take these courses to learn
[25:15] (1515.20s)
these building blocks, I feel like I'm
[25:16] (1516.80s)
getting new things that can combine to
[25:19] (1519.28s)
form kind of combinatorally or
[25:21] (1521.36s)
exponentially more software applications
[25:23] (1523.68s)
that were not possible just one or two
[25:25] (1525.92s)
years ago. So just to wrap up, this is
[25:27] (1527.76s)
my last slide. I then want to take
[25:28] (1528.96s)
questions if if y'all have any. I find
[25:31] (1531.04s)
that there are many things that matter
[25:33] (1533.20s)
for startup, not just speed. But when I
[25:35] (1535.68s)
look at the startups that AI fund is
[25:38] (1538.16s)
building, I find that the management
[25:40] (1540.16s)
team's ability to execute at speed is
[25:42] (1542.40s)
highly correlated with its odds of
[25:44] (1544.48s)
success. And some things we've learned
[25:47] (1547.12s)
to get you speed is, you know, work on
[25:49] (1549.60s)
concrete ideas. Um uh it's got to be
[25:51] (1551.92s)
good concrete ideas. I find that as a as
[25:53] (1553.76s)
executive, I'm judged on the speed and
[25:55] (1555.60s)
quality of my decisions. Both do matter,
[25:57] (1557.84s)
but speed absolutely matters. rapid
[26:00] (1560.16s)
entrying with AI coding assistance makes
[26:02] (1562.00s)
you go much faster but that shifts the
[26:04] (1564.24s)
bottleneck to getting user feedback on
[26:06] (1566.00s)
the product decisions and so having a
[26:08] (1568.00s)
portfolio of tactics to go get rapid
[26:10] (1570.48s)
feedback and if you haven't learned to
[26:12] (1572.48s)
go to coffee shop and talk to strangers
[26:14] (1574.80s)
it it's not easy but then just just be
[26:17] (1577.04s)
respectful right just be respectful of
[26:18] (1578.72s)
people that's actually very valuable
[26:20] (1580.08s)
skill for entrepreneurs to have I think
[26:22] (1582.24s)
and I think also um staying on top of
[26:24] (1584.56s)
the technology buys you speed all right
[26:27] (1587.52s)
with that let me thank Thank you very
[26:34] (1594.86s)
[Applause]
[26:40] (1600.00s)
Happy questions.
[26:41] (1601.60s)
As AI advances, do you think it's more
[26:44] (1604.00s)
important for humans to develop the
[26:47] (1607.52s)
tools or learn how to use the tools
[26:50] (1610.08s)
better? Like how can we p position
[26:52] (1612.24s)
ourselves to remain essential in a world
[26:55] (1615.20s)
where you know intelligence is becoming
[26:57] (1617.44s)
democratized? I feel like AGI has been
[27:01] (1621.44s)
overhyped and so for a long time
[27:04] (1624.16s)
there'll be a lot of things that humans
[27:05] (1625.52s)
can do that AI cannot and I think in the
[27:08] (1628.16s)
future the people that are most powerful
[27:10] (1630.40s)
are the people that can make computers
[27:13] (1633.04s)
do exactly what you want it to do and so
[27:16] (1636.24s)
I think staying on top of the tools some
[27:18] (1638.16s)
of us will build tools sometimes but
[27:19] (1639.92s)
there were a lot of other tools others
[27:21] (1641.36s)
will build that we can just use but so
[27:23] (1643.44s)
people that know how to use AI to get
[27:25] (1645.68s)
computers to do what you want it to do
[27:27] (1647.28s)
will be much more powerful, not worry
[27:28] (1648.88s)
about people running out of things to
[27:30] (1650.24s)
do, but um people that can use AI will
[27:32] (1652.48s)
be much more powerful than people that
[27:33] (1653.84s)
don't.
[27:34] (1654.48s)
Hey, so well first of all uh thank you
[27:36] (1656.64s)
so much. I have a huge respect for you
[27:38] (1658.80s)
and I think that you are true
[27:40] (1660.32s)
inspiration for a lot of us. My question
[27:42] (1662.72s)
is about uh the future of compute. So as
[27:45] (1665.44s)
we move towards uh more powerful more
[27:49] (1669.04s)
powerful AI, where do you think that
[27:51] (1671.52s)
comput is heading? I mean we see people
[27:53] (1673.20s)
saying let's ship GPUs to space. Some
[27:56] (1676.16s)
people talking about nuclear power data
[27:58] (1678.00s)
centers. What do you think about it?
[27:59] (1679.60s)
There's something I'm debating what I
[28:00] (1680.56s)
wanted to say in response to the last
[28:01] (1681.76s)
question about kind of AGI about maybe
[28:03] (1683.92s)
I'll answer this and a little bit the
[28:05] (1685.36s)
last question. So it turns out there's
[28:07] (1687.36s)
one framework you can use for deciding
[28:10] (1690.24s)
what's hype and what's not hype. I think
[28:11] (1691.76s)
over the last two years there's been a
[28:14] (1694.08s)
handful of companies that um hyped up
[28:17] (1697.52s)
certain things for promotional PR
[28:20] (1700.88s)
fundraising influence purposes. And
[28:24] (1704.24s)
because AI was so new, um, handful of
[28:27] (1707.52s)
companies got away with saying almost
[28:29] (1709.36s)
anything without anyone fact-checking
[28:31] (1711.68s)
them because the technology was not
[28:33] (1713.04s)
understood. So, one of my mental filters
[28:36] (1716.00s)
is there's certain hype narratives that
[28:38] (1718.24s)
make these businesses look more powerful
[28:40] (1720.00s)
that's been amplified. Um, and so, for
[28:42] (1722.08s)
example, this idea that um, AI is so
[28:45] (1725.76s)
powerful, we might accidentally lead to
[28:48] (1728.48s)
human extinction. That's just
[28:50] (1730.40s)
ridiculous. But it is a hype narrative
[28:53] (1733.12s)
that made certain businesses look more
[28:54] (1734.80s)
powerful and it got you know ramped up
[28:56] (1736.96s)
and actually helped certain businesses
[28:58] (1738.24s)
fundraising goals. AI is so powerful
[29:01] (1741.20s)
soon no one will even have a job
[29:02] (1742.64s)
anymore. Just not true, right? But again
[29:05] (1745.12s)
that made these businesses look more
[29:06] (1746.72s)
powerful got hyped up or we are so
[29:09] (1749.76s)
powerful so when the hype narrative
[29:12] (1752.64s)
we're so powerful that by training a new
[29:15] (1755.12s)
model we will casually wipe out
[29:16] (1756.96s)
thousands of startups. That's just not
[29:18] (1758.96s)
true. Yes, Jasper ran into trouble.
[29:21] (1761.76s)
Small number of companies got wiped out.
[29:23] (1763.36s)
But it's not that easy to casually wipe
[29:25] (1765.20s)
out thousands of startups. AI needs so
[29:28] (1768.16s)
much electricity. Only nuclear power is
[29:30] (1770.64s)
good enough for that. You know, that
[29:31] (1771.84s)
wind solar stuff not that's just not
[29:33] (1773.76s)
true. So, I think a lot of this um GPUs
[29:36] (1776.40s)
in space, you know, I don't know. It's
[29:39] (1779.36s)
like um go for it. I think we have a lot
[29:42] (1782.48s)
of room to run still for terrestrial
[29:44] (1784.16s)
GPUs. Uh yeah, but but I think uh uh
[29:46] (1786.88s)
some of these hype narratives are have
[29:49] (1789.04s)
been amplified that that I think uh are
[29:51] (1791.36s)
a distortion of what what actually will
[29:53] (1793.76s)
be done.
[29:54] (1794.64s)
There's a lot of hype in um AI and how
[29:58] (1798.00s)
and nobody's really certain about how
[30:00] (1800.08s)
we're going to be building the future
[30:01] (1801.84s)
with it. But what are some of the most
[30:03] (1803.84s)
dangerous biases or overhyped narratives
[30:07] (1807.36s)
that you've seen people talk about or
[30:10] (1810.80s)
get uh poisoned by that they end up
[30:13] (1813.44s)
running with that we should try to avoid
[30:16] (1816.32s)
or be more aware of and allow us to have
[30:18] (1818.64s)
a more realistic view as we are building
[30:22] (1822.16s)
this future.
[30:23] (1823.36s)
So I think the dangerous AI narrative
[30:25] (1825.84s)
has been overhyped. Uh AI is a fantastic
[30:27] (1827.92s)
tool, but like any other powerful tool
[30:31] (1831.04s)
like electricity, lots of ways to use it
[30:33] (1833.44s)
for beneficial purposes. Also some ways
[30:35] (1835.44s)
to use it in harmful ways. I find myself
[30:37] (1837.76s)
not using the term AI safety that much.
[30:41] (1841.04s)
Um not because I think we we should
[30:42] (1842.96s)
build dangerous things, but because I
[30:44] (1844.32s)
think safety is not a function of
[30:46] (1846.32s)
technology, it's a function of how we
[30:48] (1848.08s)
apply it. So like electric motor you
[30:50] (1850.88s)
know you can't the maker of electric
[30:53] (1853.20s)
motor can't guarantee that no one will
[30:55] (1855.68s)
ever use it from unsafe downstream toss
[30:57] (1857.84s)
like use electric motor can be used to
[30:59] (1859.84s)
build a Dallas machine electric vehicle
[31:02] (1862.16s)
can be used to build a smart bomb but
[31:04] (1864.00s)
the electric motor manufacturer can't
[31:06] (1866.08s)
control how be used downstream. So
[31:08] (1868.56s)
safety is not a function of the electric
[31:10] (1870.80s)
motor as a function of how you apply it
[31:13] (1873.12s)
and I think the same thing for AI. AI is
[31:15] (1875.04s)
neither safe nor unsafe. It is how you
[31:17] (1877.28s)
apply it that makes it safe or unsafe.
[31:19] (1879.44s)
So instead of thinking about AI safety,
[31:21] (1881.44s)
I often think about responsible AI
[31:23] (1883.84s)
because it is how we use it responsibly
[31:26] (1886.32s)
hopefully or irresponsibly that
[31:28] (1888.32s)
determines whether or not what we build
[31:30] (1890.00s)
with AI technology ends up being harmful
[31:32] (1892.16s)
or beneficial. And I feel like sometimes
[31:34] (1894.88s)
that the really weird corner cases that
[31:36] (1896.80s)
get hyped up in the news. I think just
[31:38] (1898.64s)
one or two days ago there was a Wall
[31:40] (1900.64s)
Street Journal article about AI losing
[31:43] (1903.60s)
control of AI or something. And I feel
[31:45] (1905.68s)
like that article took uh corner case
[31:49] (1909.12s)
experiments run in a lab and you know
[31:51] (1911.68s)
sensationalized it in a way that I think
[31:53] (1913.52s)
was really disproportionate relative to
[31:55] (1915.60s)
the lab experiment that was being run
[31:57] (1917.44s)
and unfortunately technology is hard
[31:59] (1919.92s)
enough to understand that many people
[32:01] (1921.92s)
don't know better and so these hype
[32:03] (1923.52s)
narratives do keep on getting amplified
[32:06] (1926.32s)
um and I feel like this has been used as
[32:07] (1927.84s)
a weapon against open source software as
[32:09] (1929.60s)
well right which is really unfortunate.
[32:11] (1931.60s)
Thank you for your work. I think your
[32:13] (1933.36s)
impact is remarkable. Uh my question is
[32:17] (1937.20s)
um as aspiring founders, how should we
[32:20] (1940.56s)
be thinking about business in the world
[32:23] (1943.68s)
where anything can be disrupted in a
[32:25] (1945.60s)
day? Whatever great mode, product or
[32:28] (1948.56s)
feature you have can be replicated with
[32:30] (1950.96s)
VIP code and competitors in basically
[32:33] (1953.28s)
hours.
[32:34] (1954.40s)
It turns out when you start a business,
[32:36] (1956.64s)
there are a lot of things to worry
[32:37] (1957.92s)
about. The number thing I worry about is
[32:41] (1961.36s)
uh are you building a product that users
[32:43] (1963.44s)
love? Um it turns out that when you
[32:46] (1966.08s)
build a business there are lots of
[32:47] (1967.44s)
things to think about the go to market
[32:49] (1969.12s)
channel competitors technology mode all
[32:51] (1971.60s)
that is important but if I were to have
[32:53] (1973.92s)
a singular focus on one thing it is are
[32:56] (1976.32s)
you building a product that users really
[32:58] (1978.24s)
want until you solve that you know is
[33:00] (1980.96s)
very difficult to build a valuable
[33:02] (1982.48s)
business. After you solve that the other
[33:05] (1985.12s)
questions do come to play. Uh, do you
[33:07] (1987.28s)
have a channel to get to customers? What
[33:09] (1989.04s)
is pricing long-term? What is your moat?
[33:11] (1991.76s)
I find that moes tend to be overhyped.
[33:14] (1994.72s)
Actually, I find that more businesses
[33:16] (1996.32s)
tend to start off with a product and
[33:18] (1998.32s)
then evolve eventually into a moat. But
[33:21] (2001.36s)
consumer products rand is somewhat more
[33:24] (2004.40s)
defensible. Um, and if you have a lot of
[33:26] (2006.72s)
momentum, it becomes harder to catch
[33:28] (2008.32s)
you. But enterprise products sometimes
[33:30] (2010.64s)
if you have a uh maybe mo is more of a
[33:33] (2013.44s)
consideration if they're channels that
[33:35] (2015.44s)
are hard to get into enterprises. So I
[33:37] (2017.92s)
think um sorry when when AI fund looks
[33:40] (2020.80s)
at businesses we actually wind up doing
[33:43] (2023.28s)
a fairly complex analysis of these
[33:45] (2025.20s)
factors and writing a you know two to
[33:47] (2027.36s)
six page narrative memo to analyze it
[33:49] (2029.60s)
before we decide whether or not to
[33:51] (2031.68s)
proceed it or not. And and I think um uh
[33:55] (2035.20s)
all of these things are important, but I
[33:57] (2037.52s)
feel like at this moment in time, the
[34:00] (2040.16s)
number of opportunities, meaning the
[34:01] (2041.92s)
amount of stuff that is possible that no
[34:04] (2044.16s)
one's built yet in the world, seems much
[34:06] (2046.40s)
greater than the number of people with
[34:08] (2048.32s)
the skill to build them. So definitely
[34:10] (2050.48s)
at the application layer, it feels like
[34:12] (2052.56s)
there's a lot of white space for new
[34:14] (2054.48s)
things you can build that no one else
[34:16] (2056.32s)
seems to be working on. And I would say,
[34:19] (2059.28s)
you know, focus on building a product
[34:20] (2060.96s)
that people want, that people love. Um,
[34:23] (2063.12s)
and then figure out the rest of it along
[34:25] (2065.68s)
the way. Although this important figure
[34:27] (2067.12s)
along the way.
[34:28] (2068.24s)
Uh, hi professor. Uh, thanks for your
[34:30] (2070.32s)
wonderful speech. Uh, I am a
[34:32] (2072.64s)
Andagramress researcher from Stanford
[34:34] (2074.48s)
and I think your uh, metaphor in your
[34:37] (2077.20s)
speech is very interesting. You said the
[34:39] (2079.68s)
uh, current AI tools are like bricks and
[34:42] (2082.56s)
are be uh, and can be built upon
[34:44] (2084.80s)
accumulation. However, so far it is
[34:47] (2087.52s)
difficult to see the accumulative
[34:49] (2089.28s)
functional expansion of the uh
[34:51] (2091.52s)
integration of AI tools because they
[34:53] (2093.36s)
often align on the stacking of functions
[34:55] (2095.68s)
based on uh intent distribution and are
[34:59] (2099.04s)
accompanied by dynamic problems of
[35:01] (2101.36s)
tokens and time overhead. So um which is
[35:04] (2104.56s)
uh which is different from static
[35:06] (2106.24s)
engineering. So what do you think will
[35:08] (2108.32s)
be the perspective of a possible agent 2
[35:10] (2110.64s)
accumulation effect in the future? But
[35:13] (2113.12s)
hey just just some quick remarks to that
[35:15] (2115.04s)
right you mentioned agent uh OM token
[35:17] (2117.60s)
cost my most common advice to developers
[35:21] (2121.12s)
is to first approximation just don't
[35:23] (2123.44s)
worry about how much tokens cost only a
[35:25] (2125.76s)
small number of startups are lucky
[35:27] (2127.28s)
enough to have users use so much of your
[35:30] (2130.08s)
product that the cost of tokens becomes
[35:32] (2132.08s)
a problem it can become a problem I've
[35:34] (2134.56s)
definitely been on a bunch of teams
[35:36] (2136.00s)
where the cost you know users like our
[35:37] (2137.84s)
product and we started to look at our
[35:39] (2139.92s)
right geni uh uh bills and it was
[35:42] (2142.48s)
definitely climbing in a way that really
[35:44] (2144.32s)
became a problem. But it's actually
[35:46] (2146.24s)
really difficult to get to a point where
[35:48] (2148.32s)
your token usage costs are a problem.
[35:50] (2150.72s)
And for the teams I'm on where we were
[35:52] (2152.80s)
lucky enough that users made our token
[35:55] (2155.36s)
cause a problem, we often had engine
[35:57] (2157.92s)
solutions to then bend the cursor and
[36:00] (2160.08s)
bring it back down through prompting
[36:01] (2161.68s)
fine-tuning USDs by optimize or
[36:03] (2163.92s)
whatever. And then what I'm seeing is
[36:06] (2166.24s)
that I'm seeing a lot of agentic
[36:08] (2168.16s)
workflows that actually integrate a lot
[36:10] (2170.00s)
of different steps. So for example, if
[36:12] (2172.08s)
you build a customer service chatbot,
[36:14] (2174.32s)
we'll often have to use prompting, maybe
[36:16] (2176.72s)
optimize some of the results in DSPI,
[36:18] (2178.64s)
build evals, build guard rails, maybe
[36:20] (2180.64s)
the customer service chatbot needs rag
[36:22] (2182.40s)
up part of the way to get information to
[36:23] (2183.84s)
feedback to the user. So I actually do
[36:25] (2185.92s)
see these things grow. But one tip for
[36:28] (2188.88s)
many of you as well is I will often
[36:31] (2191.28s)
architect my software to make switching
[36:34] (2194.40s)
between different building block
[36:35] (2195.84s)
providers relatively easy. So for
[36:38] (2198.40s)
example, um have a lot of products that
[36:40] (2200.88s)
build on top of OM, but sometimes you
[36:43] (2203.12s)
point to a specific product and ask me
[36:45] (2205.28s)
which OM are we using? I honestly don't
[36:47] (2207.92s)
know because we built up evals and when
[36:50] (2210.80s)
there's a new model that's released,
[36:52] (2212.64s)
we'll quickly run evals to see if the
[36:54] (2214.48s)
new model is better than the old one.
[36:56] (2216.16s)
And then you'll just switch to the new
[36:57] (2217.84s)
model if the new model does better on
[36:59] (2219.44s)
evals. And so the model we use week by
[37:02] (2222.00s)
week, you know, sometimes our engines
[37:03] (2223.68s)
will change it without even bothering to
[37:05] (2225.36s)
tell me because the eval show the new
[37:06] (2226.96s)
model works better. So it turns out that
[37:09] (2229.20s)
switching cost for foundation models is
[37:12] (2232.00s)
relatively low and we often architect
[37:14] (2234.32s)
our software. Oh, AI suite is open
[37:16] (2236.40s)
sourcing that my friends and I worked on
[37:18] (2238.00s)
to make switching easier. Um, switching
[37:20] (2240.24s)
cost for the orchestration platforms is
[37:22] (2242.64s)
a little bit harder. Uh but I find that
[37:25] (2245.52s)
preserving that flexibility in your
[37:27] (2247.68s)
choice of building blocks often let you
[37:29] (2249.44s)
go faster even as you're building more
[37:31] (2251.68s)
and more things on top of each other. Um
[37:33] (2253.68s)
so hope that helps.
[37:34] (2254.48s)
Thank you so much.
[37:35] (2255.44s)
In the world of education in AI, there
[37:37] (2257.20s)
are two paradigms mostly. So one is AI
[37:39] (2259.92s)
can make teachers more productive. Uh
[37:41] (2261.36s)
automating grading and automating
[37:42] (2262.88s)
homeworks. But another school of thought
[37:44] (2264.48s)
is that there'll be personal tutors for
[37:46] (2266.96s)
every student. So every student can have
[37:48] (2268.96s)
one tutor that gets feedback from an AI
[37:51] (2271.68s)
and gets personal questions from them.
[37:53] (2273.36s)
So how do you see these two paradigms
[37:54] (2274.80s)
converge and how would education look
[37:56] (2276.24s)
like in the next five years?
[37:57] (2277.36s)
I think everyone feels like a change is
[37:59] (2279.12s)
coming in edtech but I don't think the
[38:00] (2280.96s)
disruption is here yet. I think a lot of
[38:02] (2282.72s)
people are experimenting with different
[38:03] (2283.92s)
things. So you know Corsera has Corsera
[38:06] (2286.08s)
coach which actually works really well.
[38:07] (2287.84s)
Um deep learn is more focused on
[38:10] (2290.00s)
teaching AI also has some built-in chat
[38:12] (2292.00s)
bots. Um a lot of teams experiment of
[38:14] (2294.24s)
autograding. Oh there's an avatar with
[38:16] (2296.48s)
me on the deep learn website you can
[38:18] (2298.24s)
talk to if you want. Uh deep learn.ai.
[38:21] (2301.84s)
And then I think for some things like
[38:23] (2303.12s)
language learning with you know speak
[38:25] (2305.44s)
Dolingo that has become clearer some of
[38:28] (2308.40s)
the ways AI would transform it for the
[38:30] (2310.48s)
broader educational landscape the exact
[38:32] (2312.88s)
ways that AI would transform it I see a
[38:35] (2315.28s)
lot of experimentation I think what key
[38:37] (2317.76s)
learning which I've been doing some work
[38:39] (2319.20s)
with is doing is is very promising for
[38:41] (2321.20s)
K12 education but I think uh what I'm
[38:44] (2324.08s)
seeing is frankly tons of
[38:45] (2325.20s)
experimentation but the final end state
[38:47] (2327.44s)
is still not clear. I do think education
[38:49] (2329.84s)
will be hyperpersonalized. Uh but that
[38:52] (2332.40s)
workflow is an avatar, is a text
[38:54] (2334.48s)
chatbot, what's the workflow? I think um
[38:57] (2337.04s)
I feel like the hype from a couple years
[38:58] (2338.64s)
ago that with AGI soon and it will be
[39:00] (2340.88s)
all so easy. That was hype. The reality
[39:03] (2343.52s)
is work is complex, right? teachers,
[39:07] (2347.36s)
students, people do really complex
[39:09] (2349.68s)
workflows and for the next decade we'll
[39:13] (2353.12s)
be looking at the work that needs to be
[39:14] (2354.96s)
done and figuring out how to map it to
[39:17] (2357.28s)
agentic workflows and education is one
[39:20] (2360.16s)
of the sectors where this mapping is
[39:22] (2362.80s)
still underway but it's not yet mature
[39:25] (2365.04s)
enough to the point where the end state
[39:27] (2367.20s)
is clear. So I think I think we should
[39:29] (2369.84s)
all yeah just keep working on it.
[39:31] (2371.28s)
All right. All right. Thank you so much
[39:32] (2372.32s)
Andrew.
[39:32] (2372.88s)
Thank you. Uh hey my question is I think
[39:35] (2375.60s)
AI uh it has a lot of great potential
[39:37] (2377.44s)
for good but there's also a lot of
[39:38] (2378.88s)
potential for bad consequences as well
[39:41] (2381.04s)
such as exacerbating economic inequality
[39:43] (2383.44s)
and things like that and I think a lot
[39:44] (2384.88s)
of our startups here while they'll be
[39:46] (2386.56s)
doing a lot of great things will also be
[39:48] (2388.96s)
you know just by virtue of their product
[39:50] (2390.56s)
be contributing to some of those
[39:51] (2391.84s)
negative consequences. So I was curious
[39:53] (2393.92s)
how do you think you know us as AI
[39:56] (2396.24s)
builders should kind of balance our uh
[39:59] (2399.52s)
product building with also the potential
[40:01] (2401.76s)
societal downsides of some AI products
[40:04] (2404.32s)
and essentially how can we uh both move
[40:06] (2406.64s)
fast and be responsible as you mentioned
[40:08] (2408.72s)
in your talk
[40:09] (2409.84s)
look in your heart and if fundamentally
[40:12] (2412.16s)
what you're building if you don't think
[40:13] (2413.68s)
it'll make people at large better off
[40:17] (2417.04s)
don't do it right I I know it sounds
[40:18] (2418.80s)
simple but actually really hard to do in
[40:20] (2420.24s)
the moment but AI fund we've killed
[40:22] (2422.08s)
multiple projects projects not on
[40:23] (2423.76s)
financial grounds but on ethical grounds
[40:25] (2425.60s)
where there are multiple projects we
[40:26] (2426.88s)
looked at the economic case is very
[40:28] (2428.88s)
solid but we said you know what we don't
[40:31] (2431.20s)
want this exist in the world and we just
[40:32] (2432.56s)
killed it on that basis so I hope more
[40:34] (2434.80s)
people will do that and then I worry
[40:37] (2437.92s)
about uh bring everyone with us one
[40:40] (2440.64s)
thing I'm seeing is um people in all
[40:42] (2442.56s)
sorts of job roles that are not
[40:44] (2444.64s)
engineering are much more productive if
[40:46] (2446.64s)
they know AI than if they don't and so
[40:49] (2449.36s)
for example on my marketing team my
[40:51] (2451.52s)
marketers they know how the code.
[40:52] (2452.96s)
Frankly, they they were running circles
[40:55] (2455.04s)
around the ones that don't. So, then
[40:56] (2456.56s)
everyone learned to code and then they
[40:58] (2458.16s)
got better. But I feel like um trying to
[41:00] (2460.64s)
bring everyone with us to make sure
[41:02] (2462.16s)
everyone is empowered to build with AI.
[41:04] (2464.32s)
That'll be an important part of what all
[41:06] (2466.00s)
of us do, I think.
[41:07] (2467.52s)
Um I'm one of your big fans and thank
[41:09] (2469.60s)
you for your online courses. Your
[41:11] (2471.68s)
courses make the deep learning like uh
[41:13] (2473.84s)
much more accessible to the world. And
[41:16] (2476.08s)
my question is also about education. uh
[41:18] (2478.80s)
as AI becomes more powerful and
[41:20] (2480.88s)
widespread, there seems to be a growing
[41:23] (2483.04s)
gap between what can actually do and
[41:25] (2485.60s)
what people perceive it. So what do you
[41:28] (2488.56s)
think about like is it important to
[41:31] (2491.36s)
educate the general public about deep
[41:33] (2493.92s)
learning stuff and not only like uh
[41:36] (2496.32s)
educate those technical people and make
[41:38] (2498.96s)
people understand more what really uh
[41:41] (2501.20s)
what AI really do and how it works.
[41:43] (2503.84s)
I think that knowledge will diffuse deep
[41:45] (2505.44s)
learn AI we want to empower everyone to
[41:47] (2507.20s)
build with AI. So we're working on it.
[41:48] (2508.96s)
Many of us work on it. I'll just tell
[41:50] (2510.88s)
you what I think is the main d I think
[41:52] (2512.88s)
there are maybe two dangers. One is if
[41:55] (2515.04s)
you don't bring people with us fast
[41:56] (2516.48s)
enough, I hope we'll solve that. There's
[41:58] (2518.48s)
one other danger which is um it turns
[42:00] (2520.96s)
out that if you look at the mobile
[42:02] (2522.40s)
ecosystem, mobile phones, it's actually
[42:04] (2524.72s)
not that interesting. And one of the
[42:06] (2526.32s)
reasons is there are two gatekeepers,
[42:08] (2528.00s)
Android and iOS. And unless they let you
[42:10] (2530.40s)
do certain things, you're not allowed to
[42:12] (2532.40s)
try certain things on mobile. And I
[42:14] (2534.16s)
think this, you know, hampers
[42:15] (2535.60s)
innovators. These dangers of AI have
[42:18] (2538.48s)
been used by certain businesses. They're
[42:21] (2541.20s)
trying to shut down open source because
[42:23] (2543.12s)
a number of businesses that love to be a
[42:24] (2544.80s)
gatekeeper to large scale foundation
[42:27] (2547.12s)
models. So I think hyping up dangers,
[42:30] (2550.08s)
supposed false dangers of AI in order to
[42:32] (2552.88s)
get regulators to pass laws like the
[42:35] (2555.44s)
proposed SP 1047 in California, which
[42:37] (2557.84s)
thank goodness we shut down, would have
[42:40] (2560.08s)
put in place really burdensome regry
[42:42] (2562.16s)
requirements that don't make anyone
[42:43] (2563.92s)
safer, but would make it really
[42:45] (2565.60s)
difficult for TS to release open source
[42:47] (2567.28s)
and open weight software. So one of the
[42:49] (2569.68s)
dangers to inequality as well is if
[42:52] (2572.16s)
these regulatory you know awful regry
[42:54] (2574.96s)
approaches and I've been in the room
[42:56] (2576.56s)
where some of these businesses said
[42:58] (2578.24s)
stuff to regulators that was just not
[43:00] (2580.16s)
true. So I think that um some of these
[43:03] (2583.04s)
arguments the danger is if these
[43:04] (2584.72s)
regulatory proposals succeed and end up
[43:07] (2587.36s)
siphoning regulations leaving us with a
[43:09] (2589.36s)
small number of gatekeepers where
[43:11] (2591.28s)
everyone needs the permission of a small
[43:13] (2593.36s)
number of companies to fine-tune the
[43:15] (2595.04s)
model prompt in a certain way that's
[43:17] (2597.36s)
what will cipher innovation and prevent
[43:19] (2599.04s)
the diffusion of this information to let
[43:22] (2602.08s)
lots of startups you know build whatever
[43:24] (2604.40s)
they want responsibly but the freedom to
[43:26] (2606.16s)
innovate so I think so long as we um
[43:29] (2609.12s)
prevent this line of attack on open
[43:31] (2611.04s)
source open weight models from
[43:32] (2612.88s)
succeeding and we we've made good
[43:34] (2614.56s)
progress but the threat is still there
[43:36] (2616.64s)
then I think eventually we'll get to the
[43:38] (2618.40s)
diffusion of knowledge and we can
[43:40] (2620.24s)
hopefully then bring everyone with us
[43:42] (2622.16s)
but this fight to protect open source
[43:44] (2624.32s)
we've been winning but the fight is
[43:45] (2625.84s)
still on and we still have to keep up
[43:47] (2627.92s)
that work to to protect open source
[43:50] (2630.72s)
thank you all very much it's wonderful