[00:00] (0.16s)
You interviewed at 46 companies. What
[00:03] (3.12s)
did you learn about what the market is
[00:05] (5.04s)
like, what interviewing is like, what
[00:06] (6.80s)
the whole scene is in terms of the
[00:09] (9.28s)
space. There are the product companies,
[00:11] (11.20s)
infrastructure companies, and the model
[00:12] (12.88s)
companies. I found it helpful to put
[00:14] (14.96s)
companies in each category and figure
[00:17] (17.52s)
out which segment you're most excited
[00:19] (19.60s)
about to help narrow down the options
[00:21] (21.84s)
given that there's so many AI companies
[00:23] (23.44s)
right now. Product companies are the
[00:24] (24.96s)
companies building on top of the model.
[00:26] (26.48s)
Here I think of cursor, kodium, hebia.
[00:28] (28.88s)
Infrastructure companies are the
[00:30] (30.16s)
companies building the tools to help AI
[00:33] (33.04s)
product companies effectively use LLMs.
[00:35] (35.84s)
So whole suite of these there are the
[00:37] (37.44s)
inference providers like modal fireworks
[00:39] (39.44s)
together. Vector database companies like
[00:41] (41.36s)
pine cone, chromadb, evate, eval
[00:44] (44.08s)
observability tools like brain trust,
[00:45] (45.92s)
arise, galileo and a whole other suite
[00:48] (48.40s)
of products. And then there's the model
[00:50] (50.40s)
companies which are the base of the
[00:51] (51.68s)
ecosystem building the intelligence. You
[00:53] (53.92s)
have the big tech companies like Google,
[00:55] (55.92s)
Meta building models and then you also
[00:57] (57.92s)
have startups like or not startups you
[00:59] (59.76s)
have other smaller companies like OpenAI
[01:01] (61.96s)
Anthropic building models as well. So
[01:04] (64.40s)
that's how I kind of think about it. So
[01:05] (65.68s)
for me in trying to be more focused in
[01:07] (67.92s)
my search I decided to focus on model
[01:10] (70.08s)
and infrastructure companies because I
[01:11] (71.92s)
wanted to keep getting breath in what I
[01:14] (74.16s)
was doing and I felt like the product
[01:15] (75.44s)
companies were too similar to my
[01:16] (76.88s)
experience at KOD which was phenomenal
[01:18] (78.64s)
but I wanted to keep growing and that
[01:20] (80.24s)
definitely the trade-off was that it's a
[01:21] (81.76s)
bit more of an uphill battle because the
[01:23] (83.28s)
work that I had done was not as relevant
[01:25] (85.36s)
to model or infrastructure companies.
[01:27] (87.60s)
What is AI engineering and what does it
[01:29] (89.12s)
take to get hired as an AI engineer?
[01:31] (91.36s)
John Leo is a software engineer turned
[01:33] (93.28s)
AI engineer with four years of
[01:34] (94.72s)
experience but already a very impressive
[01:36] (96.84s)
background. During college, she joined a
[01:38] (98.96s)
university incubator where she shipped
[01:40] (100.32s)
four mobile apps in production to paying
[01:42] (102.08s)
customers. She then interned at Google
[01:44] (104.08s)
and Microsoft, joined Kod as a software
[01:46] (106.24s)
engineer, became one of the first AI
[01:47] (107.92s)
engineers at the company and then
[01:49] (109.60s)
interviewed with 46 different AI
[01:51] (111.44s)
startups and now works at OpenAI. In
[01:54] (114.32s)
today's conversation, we cover Jambi's
[01:56] (116.80s)
decision-making when she decided to join
[01:58] (118.32s)
a startup after graduating when she
[02:00] (120.24s)
already had returned offers from big
[02:01] (121.76s)
tech companies like Google and
[02:02] (122.96s)
Microsoft. How Jambi became one of the
[02:05] (125.04s)
first AI engineers at KOD despite being
[02:07] (127.20s)
told no when she first volunteered to
[02:09] (129.28s)
join KOD's in-house AI team. What Jambi
[02:11] (131.76s)
works on at OpenAI and why she thinks
[02:13] (133.44s)
OpenAI moves so fast and many more
[02:16] (136.00s)
topics. If you're interested in AI
[02:18] (138.16s)
engineering or how to make the
[02:19] (139.52s)
transition from software engineer to AI
[02:21] (141.12s)
engineer, this conversation has lots of
[02:22] (142.80s)
relevant tips and observations coming
[02:24] (144.48s)
from Jambi. If you enjoy the show,
[02:26] (146.72s)
please subscribe to the podcast on any
[02:28] (148.08s)
podcast platform and on YouTube. So,
[02:30] (150.24s)
Jambi, welcome to the podcast. Thanks
[02:32] (152.72s)
for having me, Ger. You were in a a good
[02:34] (154.72s)
college in in Dartmouth, but you you
[02:36] (156.72s)
then got an interest at at Google. Now,
[02:38] (158.48s)
not everyone working at Dartmouth can
[02:40] (160.00s)
get into uh a place like Google is very
[02:42] (162.40s)
competitive. How did you get that
[02:44] (164.72s)
internship and then the Microsoft
[02:45] (165.92s)
internship? what was the interview
[02:47] (167.20s)
process like and what do you think
[02:49] (169.20s)
helped you get your step in in in the
[02:51] (171.92s)
door basically back then I didn't know
[02:54] (174.32s)
anyone at Google or Microsoft so I
[02:56] (176.80s)
applied through their portal and I
[02:58] (178.32s)
remember for university students they
[03:00] (180.00s)
asked you to write essays on why you
[03:01] (181.92s)
want to work there so I remember in
[03:03] (183.68s)
those essays talking about the things
[03:06] (186.16s)
that I had built outside of classes as
[03:08] (188.32s)
well as why I wanted to work there in
[03:10] (190.68s)
particular I was lucky to get picked up
[03:13] (193.04s)
from the stack to be honest and then
[03:15] (195.36s)
leak coded to prepare for their
[03:17] (197.04s)
interviews. So, so, so tell tell me
[03:18] (198.80s)
about like preparing for for lead code.
[03:20] (200.32s)
I mean these days it is somewhat
[03:21] (201.52s)
commonly known but there's you know two
[03:23] (203.04s)
sets of people some some engineers or
[03:24] (204.72s)
college students they roll their eyes
[03:25] (205.92s)
saying this is you know pointless it's
[03:28] (208.08s)
not the job etc. And then some people
[03:29] (209.92s)
just you know like you sounds like you
[03:32] (212.16s)
just kind of like went through it
[03:33] (213.20s)
studied it prepared it how did that go?
[03:35] (215.68s)
So for Google that was in my sophomore
[03:37] (217.52s)
year and I remember being surprised that
[03:40] (220.40s)
I even got an interview in the first
[03:41] (221.92s)
place. So, I wasn't studying actively
[03:45] (225.24s)
before I when I got the interview, I
[03:48] (228.16s)
asked a couple friends, what do I study?
[03:52] (232.08s)
I think they sent us a pamphlet of
[03:53] (233.92s)
things to look at. And in that pamphlet,
[03:56] (236.64s)
there was that green book because back
[03:59] (239.20s)
then, neat code wasn't a thing. There
[04:01] (241.44s)
wasn't blind 75. And so, that green
[04:04] (244.64s)
book, I'm forgetting what it's called,
[04:06] (246.16s)
but I remember buying that book, locking
[04:08] (248.96s)
myself in my room, cracking the coding
[04:10] (250.80s)
interview probably. just cracking the
[04:12] (252.64s)
coding interview and just reading as
[04:14] (254.32s)
many questions as I could back then.
[04:16] (256.32s)
Yeah, you have it. Yeah, cracking the
[04:18] (258.56s)
book. I I even have a version of it
[04:20] (260.96s)
which was the white book that that was
[04:22] (262.56s)
like 10 10 years ago. But so the author
[04:25] (265.60s)
of this is was actually a Google
[04:27] (267.44s)
interviewer like I think 15 or so years
[04:29] (269.68s)
ago, G Gail Lman McDonald and now she
[04:33] (273.36s)
she actually sometimes I'm not sure if
[04:35] (275.76s)
she still does it, but she used to run
[04:38] (278.32s)
training programs at companies at Uber.
[04:40] (280.32s)
she came and she ran our training
[04:43] (283.04s)
program on how to do coding interviews,
[04:46] (286.00s)
uh what kind of signals to get, how to
[04:47] (287.68s)
how to change it. So, it's actually
[04:48] (288.88s)
really nice because she she teaches the
[04:50] (290.56s)
companies on how to do it. So, then she
[04:52] (292.16s)
can update the book and and actually
[04:53] (293.76s)
have up to date of how it works at
[04:55] (295.52s)
different companies. Wow. She definitely
[04:57] (297.92s)
changed the game and that I'm not sure
[04:59] (299.76s)
how much things were written down before
[05:02] (302.08s)
that. So, I think she definitely paved
[05:03] (303.68s)
the way for Neat Code and other people
[05:05] (305.92s)
to build on top of on top of this. Yeah,
[05:08] (308.72s)
if you want to build a great product,
[05:10] (310.24s)
you have to ship quickly. But how do you
[05:12] (312.72s)
know what works? More importantly, how
[05:15] (315.20s)
do you avoid shipping things that don't
[05:17] (317.08s)
work? The answer, Statig. Static is a
[05:21] (321.20s)
unified platform for flags, analytics,
[05:23] (323.52s)
experiments, and more. Combining five
[05:25] (325.60s)
plus products into a single platform
[05:27] (327.44s)
with a unified set of data. Here's how
[05:30] (330.00s)
it works. First, Static helps you ship a
[05:32] (332.88s)
feature with a feature flag or config.
[05:35] (335.52s)
Then it measures how it's working from
[05:37] (337.84s)
alerts and errors to replays of people
[05:40] (340.00s)
using that feature to measurement of
[05:41] (341.84s)
topline impact. Then you get your
[05:44] (344.08s)
analytics, user account metrics and
[05:45] (345.92s)
dashboards to track your progress over
[05:47] (347.52s)
time, all linked to the stuff you ship.
[05:50] (350.00s)
Even better, Static is incredibly
[05:51] (351.76s)
affordable with the super generous free
[05:53] (353.84s)
tier, a starter program with $50,000 of
[05:56] (356.24s)
free credits, and custom plans to help
[05:58] (358.00s)
you consolidate your existing spend on
[05:59] (359.76s)
flags, analytics, or AB testing tools.
[06:02] (362.40s)
To get started, go to
[06:05] (365.16s)
statsick.com/pragmatic. That is
[06:08] (368.84s)
statsig.com/pragmatic. Happy building.
[06:11] (371.20s)
This episode is brought to you by Cinch,
[06:13] (373.28s)
the customer communications cloud
[06:15] (375.04s)
trusted by thousands of engineering
[06:16] (376.72s)
teams around the world. If you've ever
[06:19] (379.04s)
added messaging, voice, or email into a
[06:21] (381.12s)
product, you know the pain. Flaky
[06:23] (383.20s)
delivery and platform stack with
[06:24] (384.76s)
middlemen. Cinch is different. They run
[06:27] (387.68s)
their own network with direct carrier
[06:29] (389.36s)
connections in over 60 countries. That
[06:32] (392.08s)
means faster delivery, higher
[06:33] (393.84s)
reliability, and scale that just works.
[06:36] (396.80s)
Developers love Cinch for its single API
[06:38] (398.88s)
that covers 12 channels, including SMS,
[06:41] (401.36s)
WhatsApp, and RCS. Now is the time to
[06:44] (404.00s)
pay attention to RCS, rich communication
[06:46] (406.60s)
services. It's like SMS, but smarter.
[06:49] (409.52s)
Your brand name, logo, and verify check
[06:51] (411.60s)
mark, all inside the native messaging
[06:53] (413.44s)
app built by Google, now rolling out
[06:56] (416.00s)
with Apple and major carriers. RCS is
[06:58] (418.40s)
becoming the messaging standard. Cinch
[07:00] (420.80s)
is helping teams go live globally. Learn
[07:03] (423.68s)
more at
[07:05] (425.80s)
cinch.com/pragmatic. That is
[07:10] (430.92s)
sch.com/pragmatic. And so h how did your
[07:13] (433.60s)
internships go at at both Google and and
[07:15] (435.76s)
Microsoft? Must have been really
[07:17] (437.20s)
exciting to like get it. Google was the
[07:18] (438.72s)
first one, right? It was a phenomenal
[07:21] (441.28s)
experience. And it was exciting for a
[07:23] (443.20s)
couple reasons. First, it was just a
[07:25] (445.04s)
privilege to get access to these code
[07:26] (446.72s)
bases of places that I admired. I
[07:28] (448.56s)
remember when I was at Google, I was on
[07:30] (450.16s)
the search team and I would use Momma,
[07:32] (452.32s)
their internal tool uh to find
[07:34] (454.72s)
documents. And so I remember so many
[07:37] (457.04s)
weekends where I was just trying to find
[07:39] (459.76s)
company documentation on how the search
[07:41] (461.52s)
algorithm really works or comb through
[07:43] (463.92s)
the code beyond the code that I was
[07:45] (465.60s)
touching to get a sense of well what
[07:47] (467.44s)
makes Google tick. So from an
[07:49] (469.12s)
intellectual perspective, it was just
[07:50] (470.48s)
very exciting back then. Second, you
[07:53] (473.28s)
also learn a lot of technical things
[07:54] (474.72s)
that you don't get exposure to in
[07:56] (476.36s)
college, like how to effectively operate
[07:58] (478.80s)
in a large codebase, the importance of
[08:00] (480.40s)
writing unit tests. Now, when I look
[08:02] (482.32s)
back, it's trivial, but for a college
[08:03] (483.76s)
student back then, I remember it was
[08:05] (485.60s)
very important learnings that I really
[08:08] (488.32s)
value getting back then. And to me, my
[08:11] (491.04s)
favorite part was having access to
[08:13] (493.68s)
people that were five or 10 years ahead
[08:15] (495.44s)
of me in their career. I remember over
[08:18] (498.32s)
coffee chats asking many of them, you
[08:20] (500.56s)
know, what career advice do you have?
[08:22] (502.24s)
What are things that you loved in
[08:24] (504.32s)
college that I should do more of? And
[08:26] (506.96s)
some of the advice that I got really
[08:29] (509.20s)
paved decisions that I made. So that was
[08:32] (512.00s)
my favorite part of of the internships.
[08:35] (515.20s)
I would say in hindsight that given the
[08:37] (517.84s)
big tech and startups are such different
[08:41] (521.40s)
experiences and you learn so much at
[08:44] (524.00s)
each, it would be more educational to do
[08:47] (527.28s)
one startup internship and one big tech
[08:49] (529.52s)
internship to get a very robust overview
[08:51] (531.92s)
of what both experiences are like very
[08:53] (533.60s)
early. So like looking back now that
[08:55] (535.28s)
you've done both Google and Microsoft,
[08:56] (536.72s)
they were somewhat similarish. Is is it
[08:59] (539.36s)
safe to say? I mean at the high level,
[09:00] (540.88s)
right? Like we we know every company and
[09:02] (542.40s)
every team is different. Yes. at a high
[09:04] (544.16s)
level. What was different is I wanted my
[09:06] (546.32s)
junior year to work on operating systems
[09:08] (548.16s)
because at that point I had just taken a
[09:09] (549.76s)
computer architecture class and I loved
[09:11] (551.76s)
it and so I wanted to go deeper in the
[09:13] (553.36s)
stack. So from a technical perspective
[09:15] (555.04s)
they were very different but from an
[09:17] (557.20s)
experience of what do companies look
[09:20] (560.96s)
like and how do they work which is a
[09:22] (562.56s)
huge part of an internship that part was
[09:24] (564.40s)
similar. So what did you work on at
[09:26] (566.24s)
Microsoft? Was that OS? Yeah I was
[09:28] (568.32s)
working on OS uh specifically I was
[09:30] (570.16s)
working on the Azure OS team. It was a
[09:33] (573.64s)
product that lets you interact with
[09:36] (576.64s)
Azure blobs locally from your file
[09:38] (578.80s)
system. So it hydrates and dehydrates
[09:40] (580.56s)
those blobs. You can think of it like
[09:42] (582.40s)
Dropbox for Azure blobs. Yeah. Nice.
[09:45] (585.76s)
That is so cool. I mean both that you
[09:48] (588.48s)
decided that you want to do something a
[09:50] (590.40s)
lot less conventional, you know, like
[09:52] (592.08s)
not the usual SAS apps or web apps or
[09:53] (593.92s)
whatnot and that you were able to make
[09:55] (595.76s)
it happen. Did you express this
[09:57] (597.92s)
preference uh when when you got the
[09:59] (599.68s)
internship? Yes, I remember talking
[10:02] (602.08s)
about my computer architecture class
[10:03] (603.76s)
where we built up a computer from
[10:05] (605.84s)
transistors and conveying how mind-b
[10:08] (608.16s)
blown I was from that experience and how
[10:10] (610.48s)
I really wanted to work on operating
[10:12] (612.00s)
systems and then I was lucky that they
[10:13] (613.52s)
put me on that team. That's awesome. But
[10:15] (615.76s)
I think there's a learning here of like
[10:17] (617.28s)
you don't ask you don't get. So like
[10:19] (619.52s)
again just just I just remember when
[10:21] (621.44s)
when I was running I I set up the our
[10:23] (623.60s)
first internship at Uber in Amsterdam.
[10:25] (625.36s)
So for for that site and you know like
[10:28] (628.08s)
once we made an offer to to interns like
[10:30] (630.32s)
you go through the interview process but
[10:32] (632.00s)
I also ask people like if they have
[10:33] (633.44s)
preference and most people just do not
[10:35] (635.28s)
have preference. So there is this
[10:36] (636.64s)
interesting thing that if you do express
[10:38] (638.16s)
your preference I again worst case you
[10:40] (640.08s)
know you'll get you'll get whatever it
[10:41] (641.84s)
would have been but from the other side
[10:44] (644.32s)
a lot of people often don't speak up and
[10:46] (646.16s)
and you know the people who are at these
[10:47] (647.84s)
companies they really want to try this
[10:49] (649.60s)
win-win especially for internships. The
[10:51] (651.20s)
goal of an internship is to have a great
[10:53] (653.28s)
experience and companies would like you
[10:54] (654.96s)
to return. It goes both ways, right?
[10:56] (656.56s)
They evaluate you but you also evaluate
[10:58] (658.40s)
them. So they they will actually do it.
[11:00] (660.80s)
It's just a really nice learning like
[11:02] (662.40s)
yes you like express what you're hoping
[11:04] (664.64s)
for and it might just happen. Yeah. And
[11:07] (667.36s)
these companies have so much IP and so
[11:10] (670.08s)
much that we take for granted today but
[11:12] (672.24s)
are really hard technical problems that
[11:14] (674.32s)
they have solved. So, it's just a treat
[11:16] (676.16s)
to then go work on something that you
[11:18] (678.68s)
admire and get to actually see how that
[11:21] (681.36s)
code works. Absolutely. It's once you're
[11:24] (684.96s)
in there, like these companies are are
[11:26] (686.80s)
so amazing with how big they are,
[11:28] (688.64s)
especially as an intern, a lot of doors
[11:30] (690.24s)
are open. You can also just ask and
[11:31] (691.84s)
they'll be super happy to do. So, so
[11:34] (694.16s)
then you'd made a very interesting
[11:35] (695.60s)
decision because now you interned at
[11:37] (697.12s)
Google, you're interned at Microsoft. A
[11:38] (698.64s)
lot of people would, you know, be very a
[11:40] (700.64s)
lot of students or or new grads would be
[11:42] (702.48s)
super happy with just having one. uh as
[11:44] (704.80s)
I understand you could have returned to
[11:46] (706.32s)
either and then you made the decision to
[11:49] (709.52s)
not do that.
[11:51] (711.24s)
Why you know Google Microsoft you loved
[11:54] (714.00s)
the teams tell tell me about how how you
[11:57] (717.04s)
thought about the the next step of what
[11:58] (718.88s)
you would like to do after you graduate.
[12:01] (721.44s)
So I told you how I was having coffee
[12:03] (723.12s)
chats at Microsoft my junior internship
[12:05] (725.52s)
with a bunch of mentors mentioned that
[12:08] (728.24s)
startups are a great experience as well.
[12:10] (730.24s)
So, I started to evaluate the big tech
[12:12] (732.16s)
versus startup option, and I don't think
[12:14] (734.80s)
it's black and white. I think they're
[12:16] (736.16s)
really good reasons to go to both. The
[12:18] (738.64s)
way I saw it, the upside of going to big
[12:20] (740.56s)
tech was first you learn how to build
[12:22] (742.64s)
reliable software for scale. It's very
[12:24] (744.80s)
different to build something that works
[12:26] (746.88s)
versus build something that works when
[12:29] (749.28s)
it's swarmed with millions requests from
[12:31] (751.20s)
around the world and Reddus happens to
[12:32] (752.88s)
be down at the same time. Very different
[12:35] (755.64s)
skills. So, that was one upside.
[12:39] (759.28s)
different upside for big tech in general
[12:42] (762.08s)
was that you do get to work on more
[12:44] (764.96s)
moonshot projects that aren't making
[12:47] (767.12s)
money today. They don't have the same
[12:49] (769.44s)
existential crisis that startups do and
[12:52] (772.08s)
so they can work on things that you know
[12:54] (774.40s)
uh great ARVR research is happening back
[12:57] (777.28s)
in the day. I think Google was one of
[12:58] (778.56s)
the best places if you wanted to do AI
[13:00] (780.00s)
research. There also practical good
[13:01] (781.52s)
reasons to go to big tech. I'd get my
[13:04] (784.72s)
green card faster. I'd get paid more on
[13:07] (787.24s)
average. And the unfortunate reality, I
[13:10] (790.24s)
think, is that the role does hold more
[13:12] (792.16s)
weight. People are more excited about
[13:14] (794.24s)
hiring an L5 Google engineer versus an
[13:17] (797.04s)
L5 from a startup, especially if that
[13:19] (799.20s)
startup doesn't become very successful.
[13:22] (802.32s)
With that all said though, I think there
[13:24] (804.40s)
are great reasons to go to a startup.
[13:26] (806.56s)
And back then, this was hearsay based on
[13:28] (808.88s)
what I heard from mentors. But now,
[13:30] (810.40s)
having worked at a startup for three
[13:31] (811.76s)
years, I can confirm it's indeed true.
[13:34] (814.08s)
First, you just ship so much code,
[13:36] (816.32s)
right? They're more problems than
[13:37] (817.52s)
people. Once you get access to these 0
[13:39] (819.20s)
to1 green field problems that you
[13:42] (822.48s)
wouldn't necessarily get where at big
[13:44] (824.88s)
tech maybe where there are more people
[13:46] (826.48s)
than problems. Second is the breath of
[13:49] (829.36s)
skills and this is not just in the
[13:50] (830.80s)
software engineering space right from a
[13:52] (832.80s)
software engineering space maybe one
[13:54] (834.24s)
quarter you're working on a growth
[13:55] (835.52s)
hacking front-end feature and the next
[13:56] (836.88s)
quarter you're writing Terraform but
[13:58] (838.48s)
even in terms of the non-technical
[13:59] (839.84s)
skills you get an insight into how the
[14:01] (841.52s)
business works and you're expected to PM
[14:04] (844.08s)
your own work. So there's so much breath
[14:05] (845.52s)
over there and you just get more agency
[14:07] (847.36s)
in what you work on. You get the
[14:08] (848.64s)
opportunity to propose ideas that you
[14:10] (850.72s)
think would be impactful for the
[14:12] (852.16s)
business and go execute on it. So that
[14:14] (854.56s)
breath and learning opportunity to me
[14:16] (856.88s)
was a huge upset that got me very
[14:18] (858.80s)
excited about startups. It's just so
[14:20] (860.56s)
nice to hear you summarize this because
[14:23] (863.12s)
the reality what a lot of people do is
[14:24] (864.96s)
they go to one company or the other
[14:26] (866.64s)
either big tech or startup and then
[14:28] (868.32s)
they're there for a long time and then
[14:30] (870.96s)
one day they might switch but there's a
[14:33] (873.04s)
lot of like some cost fallacy you know
[14:34] (874.80s)
like you're used to this some people
[14:36] (876.32s)
actually after a few years they might go
[14:37] (877.76s)
back to the same type of company and so
[14:40] (880.32s)
I think there's a few there's relatively
[14:41] (881.84s)
few people who see this as you know with
[14:44] (884.72s)
such short uh and and focus time
[14:47] (887.44s)
difference to see the different upsides
[14:49] (889.52s)
like you have and as you said so sounds
[14:51] (891.36s)
like the upsides did happen. So you went
[14:52] (892.96s)
to KOD, right? Yes, I did go to KOD. And
[14:56] (896.56s)
then uh how how do things go? So you
[14:58] (898.88s)
mentioned some of the upsides. I assume
[15:00] (900.24s)
like that that that all all happened
[15:01] (901.92s)
there, but uh you know what other uh
[15:04] (904.56s)
things sounds like things sped up there
[15:06] (906.48s)
actually from a professional learning
[15:08] (908.16s)
and also career experience. Definitely.
[15:10] (910.80s)
I went there for growth and breath and I
[15:13] (913.20s)
definitely got that in terms of the
[15:14] (914.56s)
opportunities that I got to work on. So
[15:16] (916.48s)
it was a phenomenal experience and I'm
[15:18] (918.64s)
happy to dive into you know specific
[15:19] (919.92s)
work I did but overall just a phenomenal
[15:21] (921.60s)
experience but be before we do before
[15:24] (924.24s)
the podcast we talked a little bit about
[15:26] (926.68s)
how you thought about selecting a
[15:29] (929.76s)
startup cuz like you did go to KOD but
[15:31] (931.60s)
as I understand this was not just I'm
[15:33] (933.76s)
just going to you know like oh this
[15:35] (935.60s)
looks like a good startup you actually
[15:37] (937.84s)
thought about how to select a
[15:40] (940.24s)
potentially great startup that that
[15:42] (942.00s)
would have that kind of potential growth
[15:43] (943.52s)
potential. what is your mental model and
[15:46] (946.00s)
and how did you evaluate and how did you
[15:48] (948.32s)
kind of you know like kind of rank the
[15:50] (950.40s)
startups and what was your application
[15:52] (952.40s)
process? So back then I didn't have
[15:54] (954.80s)
startup experience and I also went to a
[15:56] (956.48s)
school on the east coast where not many
[15:57] (957.92s)
peers around me were going to startups.
[15:59] (959.92s)
So I very much looked for where are
[16:03] (963.04s)
places where I love the people in terms
[16:05] (965.28s)
of them being smart people I can learn
[16:07] (967.28s)
from as well as being very passionate
[16:09] (969.44s)
about the product because I think you do
[16:10] (970.96s)
your best work when you are passionate
[16:12] (972.24s)
about what you're building. So it was
[16:14] (974.56s)
definitely something where I looked for
[16:16] (976.56s)
from those two lenses. Today after
[16:18] (978.88s)
having been in Silicon Valley though for
[16:20] (980.64s)
four years, I have a more robust rubric
[16:24] (984.24s)
on what I look for. So that's definitely
[16:26] (986.16s)
evolved since then because one thing
[16:29] (989.36s)
that's become super clear after living
[16:31] (991.28s)
here is that your career growth at a
[16:34] (994.08s)
startup is very contingent on the
[16:35] (995.60s)
startup growing. So then how do you
[16:37] (997.36s)
choose which startup is going to grow?
[16:38] (998.88s)
And that's a hard question. You know
[16:40] (1000.72s)
venture capitalists spend all their time
[16:42] (1002.48s)
thinking about this. Yeah. And today
[16:45] (1005.12s)
what what is your mental model or for
[16:46] (1006.96s)
someone who is you know has a few years
[16:49] (1009.20s)
of experience a bit like yourself what
[16:52] (1012.40s)
would you advise for them on how to
[16:54] (1014.56s)
think about different categories of
[16:56] (1016.08s)
startups the kind of risk the upsides
[16:58] (1018.16s)
and so on. There are startups of all
[17:00] (1020.96s)
different sizes and the smaller you go
[17:03] (1023.60s)
the more risk there
[17:04] (1024.84s)
is. I think that's part of the game and
[17:07] (1027.12s)
that's what makes it exciting because
[17:08] (1028.32s)
you also have more upside when there's
[17:10] (1030.00s)
more risk. That being said, I feel very
[17:12] (1032.72s)
strongly that all engineers that take a
[17:14] (1034.48s)
pay cut to go to a startup should have
[17:16] (1036.08s)
an informed thesis on why they think
[17:18] (1038.40s)
that company is going to grow during
[17:20] (1040.72s)
their tenure. And how to actually assess
[17:24] (1044.64s)
growth is a hard question with no right
[17:26] (1046.88s)
answer. But my current rubric is looking
[17:30] (1050.88s)
for four things. First, high revenue and
[17:33] (1053.92s)
steep revenue growth rate.
[17:35] (1055.92s)
Second, a large market where there's
[17:38] (1058.40s)
room to
[17:39] (1059.48s)
expand. Third, loyal obsessed customers.
[17:43] (1063.20s)
And then fourth, competition. Why this
[17:45] (1065.12s)
company will win in that space. And I'm
[17:47] (1067.76s)
happy to go deeper into any of those,
[17:49] (1069.20s)
but that's at least how I think about
[17:51] (1071.68s)
assessing different startups today. And
[17:53] (1073.92s)
it's all relative because a startup that
[17:56] (1076.88s)
is premiumf will have less revenue than
[17:59] (1079.28s)
a startup that is series D 400 people.
[18:01] (1081.84s)
And then when you're like thinking about
[18:03] (1083.92s)
the the these four different things. So
[18:06] (1086.16s)
like we we'll later get to your your
[18:08] (1088.40s)
actual job search as well, but do you
[18:10] (1090.96s)
like try to find these things? So for
[18:12] (1092.40s)
example, you mentioned one thing about
[18:14] (1094.32s)
how customer obsession, right? Like how
[18:16] (1096.88s)
much customers love it? Like let's say
[18:18] (1098.48s)
you you have a or there's a startup that
[18:20] (1100.96s)
you're kind of interested in. How do you
[18:22] (1102.96s)
evaluate that? Do you kind of look it up
[18:24] (1104.48s)
yourself? Do you put in the work? Do you
[18:26] (1106.08s)
try to somehow outsource it? What what
[18:28] (1108.56s)
worked for you? Because there's no right
[18:30] (1110.88s)
answer here, I think it's really
[18:32] (1112.32s)
important to do the due diligence
[18:33] (1113.60s)
yourself because you're going to be the
[18:35] (1115.28s)
one that's responsible for your decision
[18:36] (1116.72s)
here, good or bad. How I think about
[18:38] (1118.64s)
things like customer obsession is I look
[18:41] (1121.84s)
on Reddit on YouTube to try to find real
[18:44] (1124.96s)
users. For more SAS companies where you
[18:47] (1127.60s)
may not have customers writing about the
[18:49] (1129.36s)
product online, I'd actually find
[18:51] (1131.20s)
companies that use that product and then
[18:53] (1133.12s)
go try to talk to them and understand
[18:55] (1135.04s)
from the ground what do people think
[18:57] (1137.12s)
about this product? especially if this
[18:58] (1138.80s)
is a product that I can't use myself
[19:00] (1140.56s)
because it's not for customers but for
[19:02] (1142.48s)
businesses instead. I I love it and
[19:04] (1144.24s)
again I I don't think more enough people
[19:07] (1147.44s)
do this kind of due diligence and and
[19:08] (1148.80s)
and they should you know one I guess now
[19:12] (1152.08s)
but in famous example is fast the
[19:14] (1154.32s)
one-click checkout startup where they
[19:16] (1156.32s)
recruited actually there were some ex
[19:18] (1158.40s)
Uber folks there who I I like knew to to
[19:21] (1161.44s)
some extent but a lot of people were
[19:22] (1162.96s)
recruited with with this shiny diagram
[19:24] (1164.72s)
that showed headcount growth and people
[19:28] (1168.16s)
most a lot of people did not ask about
[19:30] (1170.24s)
revenue or or when they did they were
[19:32] (1172.48s)
they were okay not hearing about it and
[19:34] (1174.16s)
even the people who worked there for a
[19:35] (1175.44s)
while they ignored it and there were
[19:36] (1176.80s)
some people who actually asked about it
[19:38] (1178.24s)
and they actually realized that
[19:39] (1179.68s)
something was off but just following
[19:41] (1181.36s)
your framework for example some people
[19:43] (1183.68s)
who are a bit more diligent could have
[19:45] (1185.36s)
avoided it same thing with customers for
[19:46] (1186.96s)
example there there were not many and
[19:50] (1190.08s)
like one learning that I I had back then
[19:52] (1192.40s)
and talking with engineers who worked
[19:53] (1193.84s)
there and got burnt they all told me I
[19:55] (1195.52s)
wish I would have done a bit more due
[19:56] (1196.80s)
diligence and not taken the CEO's word
[19:59] (1199.76s)
for it but also asked for proof
[20:02] (1202.72s)
say same thing revenue runway those kind
[20:04] (1204.88s)
of things. Yeah. I feel like you know at
[20:06] (1206.56s)
startups we're paid in equity a large
[20:09] (1209.44s)
chunk and so you're investors so you
[20:11] (1211.44s)
have the right to all this information
[20:13] (1213.12s)
and to me if a startup's not giving you
[20:14] (1214.80s)
that information that is a red flag in
[20:16] (1216.40s)
and of itself. Yeah. I I feel maybe
[20:18] (1218.40s)
people should think about that if you
[20:20] (1220.24s)
join a startup a bit like if you put in
[20:22] (1222.16s)
like a bunch of your money like a
[20:24] (1224.24s)
significant amount of your savings. And
[20:26] (1226.48s)
when I did angel investing, if if I
[20:28] (1228.40s)
didn't get information, I mean, you can
[20:29] (1229.84s)
still put it in and you can hope for the
[20:31] (1231.20s)
best, but I think that's called gambling
[20:32] (1232.88s)
to be fair. It is. And and so that's
[20:36] (1236.08s)
okay. But then just be honest with
[20:37] (1237.20s)
yourself like if I'm not getting this
[20:38] (1238.64s)
information, I am gambling my my most
[20:41] (1241.20s)
valuable time and very valuable years of
[20:43] (1243.20s)
of my life. And that's okay, right? It
[20:45] (1245.52s)
could work, but it's maybe not the smart
[20:47] (1247.12s)
way. Exactly. And as engineers, we have
[20:49] (1249.84s)
when we're recruiting, we're elite
[20:50] (1250.88s)
coding, we're doing system design. It's
[20:52] (1252.64s)
hard to carve out that time to do
[20:54] (1254.08s)
diligence. And so it's something I think
[20:55] (1255.44s)
we don't talk about enough. I will say
[20:57] (1257.04s)
that as a hiring manager or even as a
[20:59] (1259.56s)
manager when you join a company and if
[21:02] (1262.00s)
you previously done your due diligence,
[21:04] (1264.40s)
you will have a better start. People
[21:06] (1266.32s)
will remember you saying, "Oh, this is
[21:08] (1268.00s)
this person who actually cares about the
[21:10] (1270.32s)
business, cares about where it's going,
[21:11] (1271.92s)
cares about how they can contribute." So
[21:13] (1273.44s)
on day one, you're already not just seen
[21:15] (1275.28s)
as like, oh, you know, like new starter
[21:16] (1276.88s)
XYZ, but like oh, like this person has
[21:19] (1279.60s)
drive. Like I think that comes across.
[21:21] (1281.28s)
And honestly, if a if a company is not
[21:23] (1283.04s)
happy, you just trying to understand the
[21:25] (1285.12s)
business, seeing how you can fit in,
[21:26] (1286.80s)
that's probably red flag itself. Let's
[21:28] (1288.48s)
be real. Yeah, that's true. That's fair.
[21:30] (1290.48s)
So, at at KOD, you joined as a software
[21:32] (1292.32s)
engineer and you then transitioned into,
[21:35] (1295.84s)
you know, I looked at your LinkedIn to
[21:37] (1297.36s)
AI engineer. How did that happen? And
[21:39] (1299.60s)
and how did you make that happen?
[21:41] (1301.52s)
Because I it sounds like you actually
[21:43] (1303.12s)
had a lot to do with it. So, if we
[21:45] (1305.04s)
rewind to end of 2022, that was when
[21:48] (1308.08s)
chatbt came out and November. Oh yeah.
[21:51] (1311.92s)
Yeah. Big milestone. And you know, Kota
[21:54] (1314.72s)
saw the amount of love that this product
[21:56] (1316.64s)
was getting and KOD was starting an AI
[21:58] (1318.88s)
team with two engineers to build an AI
[22:01] (1321.36s)
assistant to help you build your KOD
[22:03] (1323.16s)
documents. At that time, I asked that,
[22:05] (1325.52s)
hey, I'd love to be a part of it and got
[22:07] (1327.44s)
a very polite no. So, I thought, no
[22:10] (1330.08s)
problem. I'm just going to start
[22:11] (1331.36s)
building in the space anyway in my
[22:13] (1333.12s)
nights and weekends because this
[22:14] (1334.32s)
technology is very cool.
[22:16] (1336.64s)
The first thing while I was learning was
[22:19] (1339.12s)
trying to answer to myself how does
[22:20] (1340.96s)
chatbt even work and through that went
[22:24] (1344.40s)
down a rabbit hole of self-studying the
[22:27] (1347.12s)
foundations of deep learning. So
[22:28] (1348.64s)
starting off with the very basics of
[22:30] (1350.56s)
what's a token, what's a weight, what's
[22:32] (1352.32s)
an embedding to then understanding that
[22:34] (1354.32s)
okay LLMs are just next token prediction
[22:37] (1357.60s)
going through the history of different
[22:39] (1359.20s)
architectures of how we went from RNNs
[22:41] (1361.28s)
to LSTMs and then building my way up to
[22:43] (1363.60s)
the transformer and understanding that
[22:45] (1365.12s)
okay it's positional encoding and
[22:46] (1366.88s)
attention that has allowed us to scale
[22:48] (1368.88s)
up in such a good way. What this did for
[22:51] (1371.12s)
me was to just give me some intuition of
[22:53] (1373.36s)
how this technology works which gave me
[22:55] (1375.84s)
a bit more
[22:57] (1377.08s)
confidence. After having built that
[22:59] (1379.28s)
foundation, I wrote about it in my in my
[23:01] (1381.68s)
blog and so my team was aware that I was
[23:03] (1383.36s)
working on this as well. I started to
[23:06] (1386.48s)
build on top of these models. So I went
[23:08] (1388.40s)
to a lot of hackathons. My favorite one
[23:11] (1391.20s)
was a way to learn languages while
[23:14] (1394.24s)
watching TV because that's the way that
[23:16] (1396.32s)
I learned Hindi and I wanted a way to
[23:18] (1398.08s)
practice my Mandarin and that's in that
[23:19] (1399.60s)
way. When I was building and doing
[23:23] (1403.12s)
hackathons, I got a sense of how to
[23:25] (1405.12s)
actually use these tools. So after 5
[23:27] (1407.76s)
months had passed, when I asked again to
[23:30] (1410.08s)
join the AI team, I got very much a heck
[23:32] (1412.80s)
yes, come join us. We see that you truly
[23:34] (1414.56s)
care about this because you've been
[23:35] (1415.68s)
working on it in your free time. And
[23:37] (1417.52s)
that's when the real learning started
[23:39] (1419.08s)
because hacking on top of models in your
[23:42] (1422.08s)
free time is very different from trying
[23:44] (1424.32s)
to build it for production especially
[23:46] (1426.64s)
because as engineers our role is to
[23:48] (1428.40s)
build reliable systems but you're
[23:49] (1429.92s)
dealing with stochastic models. So
[23:51] (1431.84s)
they're very much at odds at with each
[23:53] (1433.52s)
other. And when you say hackathons is
[23:55] (1435.60s)
this was these like weekend hackathons
[23:57] (1437.44s)
you know the ones that anyone can attend
[23:59] (1439.36s)
you register and and like especially
[24:00] (1440.72s)
they were popping up because of the AI
[24:03] (1443.20s)
uh you know like hype basically
[24:04] (1444.80s)
starting. Yes weekend hackathons. I also
[24:07] (1447.68s)
did so the project I was telling you
[24:10] (1450.48s)
about that was a language learning tool
[24:13] (1453.28s)
that was with an online hackathon for
[24:15] (1455.60s)
six weeks with this company called
[24:17] (1457.20s)
Buildspace. Anyone could go join and the
[24:19] (1459.60s)
way you win in this hackathon is not by
[24:21] (1461.36s)
what you build but how many users you
[24:23] (1463.52s)
get or how much revenue you're
[24:24] (1464.80s)
generating. So it's such a fun way as an
[24:27] (1467.04s)
engineer to not just build something but
[24:30] (1470.08s)
actually go and try to get people to use
[24:31] (1471.92s)
it. So it was a very fun learning
[24:33] (1473.20s)
experience and because it was all online
[24:35] (1475.76s)
they really encouraged us to build in
[24:37] (1477.72s)
public and that in and of itself was a
[24:40] (1480.16s)
great learning. I I I love it because I
[24:42] (1482.32s)
think uh a lot of times when a new
[24:45] (1485.12s)
technology comes out and you know a lot
[24:46] (1486.72s)
of engineers especially you had a day
[24:48] (1488.16s)
job and the people who have a day job
[24:49] (1489.36s)
the the biggest thing is like a I can
[24:51] (1491.44s)
build something on the side but what
[24:52] (1492.72s)
should I even build? I I mean, you know,
[24:54] (1494.48s)
like it it feels it's kind of pointless.
[24:56] (1496.48s)
Like you can do tutorials, but
[24:58] (1498.08s)
especially in this case, there's not
[24:59] (1499.36s)
many and tutorials are kind of not
[25:01] (1501.92s)
there. So, I love how you found a way to
[25:04] (1504.88s)
have a goal to enjoy it to do, you know,
[25:07] (1507.36s)
scratch your own itch as well and and
[25:10] (1510.32s)
combine it. So, like maybe these online
[25:12] (1512.56s)
hackathons or like hackathons happening
[25:14] (1514.64s)
around you, it could be a great way to
[25:16] (1516.88s)
do it. And and it sounds like it
[25:19] (1519.20s)
actually helped your professional like
[25:21] (1521.36s)
you help helped even your company and
[25:23] (1523.68s)
and and your job because now knowing how
[25:26] (1526.08s)
to use these tools was very much in
[25:28] (1528.72s)
demand. It still is but there were not
[25:30] (1530.56s)
many people who who were like as
[25:32] (1532.16s)
enthusiastic and as as self-saw. One
[25:34] (1534.64s)
thing that I learned from that
[25:36] (1536.48s)
experience was don't wait for people to
[25:38] (1538.64s)
give you the opportunity to do something
[25:40] (1540.88s)
just start working on it. I I I I love
[25:42] (1542.96s)
this. This is such a good mindset. So
[25:44] (1544.64s)
when you joined this team so technically
[25:47] (1547.52s)
did you become an AI engineer? What do
[25:49] (1549.52s)
you think even an AI engineer is? I feel
[25:51] (1551.20s)
is this kind of overloaded term. So I
[25:53] (1553.28s)
just love to hear like how you think
[25:55] (1555.04s)
about it. AI product engineer is
[25:57] (1557.36s)
building products on top of models and
[25:59] (1559.44s)
the work entails first a lot of
[26:02] (1562.84s)
experimentation of this new tool came
[26:05] (1565.44s)
out experimenting with what you can
[26:07] (1567.68s)
build to solve real customer problems
[26:09] (1569.68s)
prototyping it and then from there going
[26:12] (1572.24s)
actually building it from production. So
[26:13] (1573.76s)
at its core, it's very similar to
[26:16] (1576.56s)
software engineering. There are some
[26:18] (1578.64s)
domain specific things like learning how
[26:20] (1580.48s)
to fine-tune, learning how to write good
[26:23] (1583.28s)
prompts, learning how to host open
[26:25] (1585.20s)
source models, but in of itself, the
[26:27] (1587.76s)
foundation is very much software
[26:29] (1589.36s)
engineering. Yeah. And I guess you know
[26:31] (1591.28s)
like I guess evaluation is is still is
[26:33] (1593.68s)
also a big one. Yes, that's a great one.
[26:35] (1595.60s)
Writing get evals. Uh, and then like one
[26:38] (1598.40s)
thing that was really surprising for me
[26:39] (1599.76s)
to to learn when I talk with a friend
[26:41] (1601.52s)
who works at a startup is how their test
[26:44] (1604.48s)
suite costs money to run every time. The
[26:47] (1607.04s)
Eval suite, they're like, I don't know,
[26:48] (1608.72s)
like how many, like $50 or something
[26:50] (1610.88s)
like that. And it's like, oh, you know,
[26:53] (1613.36s)
when I run my unit test, like it costs
[26:55] (1615.60s)
time and and effort, but but it's it's
[26:57] (1617.44s)
it's free. It's just time. And now you
[26:59] (1619.44s)
actually, especially if you're using an
[27:00] (1620.72s)
API, you have this cost which is I think
[27:03] (1623.84s)
refreshing and just a good way to think
[27:06] (1626.00s)
about it and it just forces you to
[27:08] (1628.24s)
adapt. Yeah, for sure. It's very
[27:10] (1630.40s)
interesting because there's no good way
[27:12] (1632.96s)
to measure the accuracy of a
[27:15] (1635.36s)
nondeterministic model without using
[27:17] (1637.68s)
LLM. And so at COD we used to use brain
[27:20] (1640.16s)
trust and it was so interesting and how
[27:22] (1642.96s)
the model is being used to check whether
[27:26] (1646.56s)
or not it's working correctly. Yeah. As
[27:30] (1650.16s)
you were just going deeper and deeper
[27:32] (1652.00s)
into uh the the AI field like what were
[27:34] (1654.48s)
resources that that helped you? Was it
[27:36] (1656.16s)
just pure self-arning? Was it going to
[27:38] (1658.40s)
the the source of the where the papers
[27:41] (1661.60s)
are? like this is this is a really
[27:44] (1664.00s)
ongoing question because you know the
[27:45] (1665.52s)
industry is is not slowing down and
[27:46] (1666.88s)
there's not many kind of books or like
[27:49] (1669.68s)
you know static resources out there.
[27:51] (1671.60s)
Yeah, very fair because things are
[27:53] (1673.92s)
changing quickly and there aren't static
[27:55] (1675.96s)
resources at that time and I still true
[27:58] (1678.88s)
today. I found it most helpful to learn
[28:00] (1680.72s)
by just doing. So even when I was on
[28:03] (1683.12s)
this team I'd go to a lot of hackathons
[28:05] (1685.28s)
internal to KOD and external. I remember
[28:08] (1688.16s)
there was an internal hackathon at KOD
[28:11] (1691.12s)
where it happened to line up with the
[28:13] (1693.60s)
day OpenAI released function calling for
[28:15] (1695.84s)
the first time. And so our team we
[28:18] (1698.48s)
played around with the power of function
[28:19] (1699.92s)
calling which is a very important tool
[28:22] (1702.32s)
by turning natural language prompts into
[28:26] (1706.64s)
appropriately identifying what
[28:29] (1709.92s)
third-party code integration you should
[28:32] (1712.16s)
use. So for example, a user types in how
[28:35] (1715.12s)
many unread emails do I have? and it
[28:36] (1716.72s)
should appropriately pull out the Gmail
[28:38] (1718.96s)
pack or Gmail thirdparty integration
[28:40] (1720.96s)
that KOD had at that hackathon playing
[28:43] (1723.76s)
around with embeddings from Pine Cone to
[28:45] (1725.92s)
see can I more accurately pull out the
[28:48] (1728.80s)
right third party integration. So that
[28:51] (1731.04s)
was one way through internal hackathons
[28:53] (1733.04s)
but there were also external hackathons.
[28:54] (1734.88s)
I remember in SF there when Llama 3
[28:57] (1737.76s)
first came out they were hosting a
[29:00] (1740.00s)
fine-tuning hackathon. So I went the
[29:03] (1743.52s)
beginning they tell you what is
[29:04] (1744.88s)
fine-tuning, how to use it, which is
[29:06] (1746.72s)
great. Then there are a bunch of
[29:08] (1748.16s)
startups there that are building
[29:09] (1749.52s)
fine-tuning platforms. So they give you
[29:11] (1751.28s)
free credits to go fine-tune. And so
[29:14] (1754.08s)
then I remember building on top of
[29:15] (1755.52s)
replicate and fine-tuning a way to turn
[29:19] (1759.16s)
lama into KOD formulas, which is our
[29:22] (1762.24s)
equivalent of Excel formulas. So
[29:24] (1764.56s)
learning by doing to me was the most
[29:26] (1766.64s)
effective way when things are changing
[29:28] (1768.48s)
so quickly. And even though hackathons
[29:30] (1770.80s)
are the most effective, you know,
[29:32] (1772.08s)
reading blogs, some paper, Twitter to
[29:34] (1774.48s)
see what other people are doing did
[29:36] (1776.32s)
help, there are a lot of open source
[29:37] (1777.84s)
companies. I remember back in the day,
[29:39] (1779.52s)
Langchain had lovely documentation on
[29:42] (1782.00s)
how to do RAG when it was first getting
[29:43] (1783.68s)
popularized. And so reading what other
[29:46] (1786.48s)
people are doing, even though they're
[29:48] (1788.64s)
informally written, it's not a textbook,
[29:50] (1790.88s)
it's not a course, has been very
[29:52] (1792.56s)
informative as well. Nice. Well, yeah, I
[29:55] (1795.36s)
guess this is so new. You just need to
[29:57] (1797.04s)
figure out what works for you and just
[29:58] (1798.48s)
try a bunch of stuff and and see what
[29:59] (1799.76s)
sticks and also it changes, right? So
[30:01] (1801.36s)
like whatever works now, it it might not
[30:02] (1802.88s)
be as efficient later. So totally. Yeah.
[30:05] (1805.12s)
And there are books coming up. I
[30:06] (1806.40s)
remember you interviewed Chip and she
[30:07] (1807.68s)
has a lovely book on how to build as an
[30:10] (1810.32s)
AI engineer. Yeah. Yeah. She she
[30:12] (1812.16s)
actually captured a lot of the things
[30:13] (1813.44s)
that are not really changing anymore. So
[30:15] (1815.12s)
that that's that's also changing and I
[30:16] (1816.80s)
think you know we'll now see courses
[30:18] (1818.32s)
come out
[30:19] (1819.16s)
andre Karpathi is uh doing some really
[30:22] (1822.88s)
really in-depth like courses if if if
[30:25] (1825.36s)
you have the time which honestly it
[30:26] (1826.88s)
doesn't sound like a bad time investment
[30:28] (1828.40s)
to do so. So yeah yeah exactly with zero
[30:31] (1831.12s)
to hero. Yeah. So at at KOD what was
[30:32] (1832.88s)
your favorite project uh that that you
[30:34] (1834.48s)
built uh using AI tools or your favorite
[30:36] (1836.96s)
AI product? A project that's very close
[30:39] (1839.68s)
to my heart from KOD is Workspace Q&A.
[30:42] (1842.72s)
So maybe to set some context at KOD a
[30:45] (1845.60s)
very common customer complaint was that
[30:47] (1847.52s)
I have so many documents with my
[30:49] (1849.52s)
internal knowhow of compment need
[30:52] (1852.20s)
documentation but it's hard to find that
[30:54] (1854.24s)
document when I need it and about in
[30:57] (1857.04s)
November 2023 RAG was getting
[30:59] (1859.04s)
popularized retrieval augmented
[31:01] (1861.04s)
generation and it struck our team that
[31:03] (1863.20s)
we actually had all the tools in place
[31:05] (1865.92s)
to build a chatbot that would solve this
[31:08] (1868.16s)
problem. First, we had a team that had
[31:11] (1871.12s)
just redone our search index and they
[31:13] (1873.52s)
put a lot of hard work into redoing that
[31:15] (1875.20s)
search index. Second, we had the
[31:16] (1876.96s)
infrastructure in place to call LM tools
[31:19] (1879.88s)
reliably. And third, we had a chatbot
[31:22] (1882.24s)
that allowed you to in your KOD doc chat
[31:25] (1885.92s)
with an LLM. Yeah. With those three
[31:29] (1889.16s)
things, I was able to just glue them
[31:31] (1891.44s)
together in a couple days and build a
[31:33] (1893.04s)
version one of a chatbot that lets users
[31:35] (1895.04s)
ask questions about the content of their
[31:37] (1897.12s)
workspace. Oh, nice. So I put that you
[31:39] (1899.68s)
know on Slack with a loom and to my
[31:42] (1902.64s)
surprise uh our CEO Shashir started
[31:45] (1905.04s)
taking interest in this and responding
[31:46] (1906.88s)
to that thread. He saw a grander vision
[31:50] (1910.32s)
where Kod could create an enterprise
[31:52] (1912.40s)
search tool. So it's not just searching
[31:54] (1914.56s)
documents but all third party
[31:56] (1916.32s)
integrations which Kota had a lot of. So
[31:58] (1918.80s)
ideally, you know, a sales team should
[32:00] (1920.24s)
be able to come in and say, "What's my
[32:01] (1921.84s)
projected ARR for an account and it
[32:04] (1924.48s)
pulls in from your
[32:07] (1927.72s)
Salesforce integration and answers that
[32:10] (1930.00s)
question for you." So, that was
[32:11] (1931.60s)
exciting. And he basically tasked a
[32:14] (1934.64s)
couple of us to experimentally prove out
[32:16] (1936.96s)
that Kota could build this in four
[32:18] (1938.96s)
weeks. Oh, nice. And a good challenge.
[32:21] (1941.60s)
It was Yeah, it was a good daunting
[32:23] (1943.76s)
challenge. It was me, my manager, the
[32:26] (1946.00s)
CTO, a designer or and um a PM. And it
[32:30] (1950.40s)
was short deadlines and high stakes
[32:32] (1952.40s)
because this was going to be demoed to
[32:34] (1954.40s)
important people. So, it was very much
[32:36] (1956.88s)
all hands on deck. On one day's notice,
[32:39] (1959.44s)
we flew to New York to hack together and
[32:43] (1963.28s)
it was nights, weekends, whatever it
[32:45] (1965.44s)
took to make this work. It was very
[32:48] (1968.00s)
exciting time and I think a lot of
[32:50] (1970.96s)
blood, sweat, and tears behind it. But
[32:53] (1973.20s)
the TLDDR is that it did go very well
[32:55] (1975.84s)
and it became the birth of a second
[32:58] (1978.88s)
product for Kota called Kota Brain. From
[33:02] (1982.72s)
January to June of 2024, we had a much
[33:07] (1987.68s)
larger initiative where 20 people were
[33:09] (1989.60s)
now working on it and it was taking that
[33:12] (1992.00s)
version two that that small team we had
[33:14] (1994.56s)
built and making it a more robust thing
[33:16] (1996.24s)
which is a very hard challenge in and of
[33:17] (1997.84s)
itself. And the cherry on top was that
[33:20] (2000.16s)
Kodrain was presented at Snowflake Dev
[33:22] (2002.48s)
Day at the keynote. So it was just a
[33:24] (2004.32s)
very exciting time to be a part of it
[33:27] (2007.12s)
from day one and the world getting to
[33:30] (2010.56s)
see it at a large scale. Yeah. So I I'm
[33:33] (2013.60s)
just like taking notes on like how
[33:36] (2016.40s)
amazing it is that you know you joined
[33:38] (2018.08s)
KOD as a new grad with like no
[33:41] (2021.20s)
experience in AI engineering and all
[33:43] (2023.60s)
just frankly you know you had less
[33:45] (2025.36s)
experience than like a lot of the
[33:46] (2026.64s)
experienced engineers and software
[33:48] (2028.00s)
engineering. I mean just the the years
[33:49] (2029.84s)
of experience but from from the first
[33:52] (2032.32s)
day like you just kept track of the
[33:54] (2034.08s)
industry you saw this exciting thing is
[33:55] (2035.84s)
coming out Chad GP you tried it out you
[33:57] (2037.76s)
were convinced this this is this is
[33:59] (2039.36s)
going to be interesting and fun. You
[34:01] (2041.52s)
asked your your manager when Kota
[34:03] (2043.60s)
started a team to join, they said no and
[34:05] (2045.84s)
you just went and learned and in a
[34:07] (2047.84s)
matter of few months you probably
[34:09] (2049.28s)
leaprogged a lot of the people who were
[34:11] (2051.20s)
just kind of waiting or you know not not
[34:13] (2053.60s)
necessarily u like being as as active as
[34:16] (2056.88s)
you are. You you got onto this team as
[34:20] (2060.16s)
an early engineer and you know a year
[34:21] (2061.68s)
later now when 20 people were working on
[34:23] (2063.68s)
this with KOD you were still like one of
[34:25] (2065.28s)
the earlier ones. So it just like shows
[34:26] (2066.88s)
me how like what what you were saying
[34:29] (2069.60s)
not waiting for permission really pays
[34:32] (2072.64s)
off and you can just do things you can
[34:34] (2074.56s)
learn things and especially for for an
[34:36] (2076.56s)
innovative technology like AI and
[34:38] (2078.80s)
whatever we see next it's actually
[34:40] (2080.24s)
valuable like companies like a company
[34:43] (2083.04s)
will value this kind of thing because it
[34:44] (2084.88s)
helps them and they they want they
[34:46] (2086.24s)
desperately need people like like like
[34:48] (2088.48s)
you were in this case or or other folks
[34:50] (2090.48s)
who are doing similar things. What is
[34:52] (2092.32s)
really cool is that it's so new. So, it
[34:54] (2094.32s)
definitely levels the playing field
[34:56] (2096.56s)
between all sorts of seniorities because
[34:58] (2098.72s)
nobody knows what the right way is. And
[35:01] (2101.04s)
so, we're all just figuring it out
[35:02] (2102.16s)
together. And that's what makes it
[35:03] (2103.36s)
definitely more exciting. Yeah. And I
[35:05] (2105.52s)
feel there's like two things here. Like,
[35:07] (2107.12s)
if you're someone who already has some
[35:09] (2109.12s)
experience, may that be one year or 10
[35:10] (2110.88s)
years or 20 years, that experience will
[35:13] (2113.12s)
eventually be applicable once you
[35:14] (2114.64s)
understand how this works, you know, you
[35:16] (2116.48s)
can take that past experience and see
[35:17] (2117.84s)
how it applies. And if you don't have
[35:19] (2119.36s)
experience, it's actually not a bad
[35:20] (2120.96s)
thing because you're coming in with a
[35:22] (2122.40s)
fresh mind and you will probably you
[35:24] (2124.88s)
will not have some of those biases of
[35:27] (2127.04s)
you know for for example a lot of
[35:28] (2128.88s)
software engineers who have like 10 plus
[35:30] (2130.72s)
years of experience they will know uh
[35:33] (2133.04s)
who build production system that unit
[35:34] (2134.72s)
testing and automate testing is super
[35:36] (2136.16s)
efficient and a very good way to do
[35:37] (2137.60s)
stuff. Now with AI systems it's not
[35:40] (2140.72s)
necessarily the case when they're
[35:41] (2141.92s)
nondeterministic and things like for for
[35:44] (2144.24s)
large scale systems things like
[35:45] (2145.60s)
monitoring or checking of might be a
[35:47] (2147.60s)
better way. I'm, you know, I'm not sure
[35:49] (2149.04s)
which one it is, but not having that
[35:50] (2150.72s)
that bias could actually speed you up.
[35:52] (2152.72s)
So, like either way, it doesn't seem to
[35:55] (2155.20s)
be any any downside in just figuring it
[35:57] (2157.68s)
out and mastering this tool because it
[35:59] (2159.92s)
is a tool at the end of the day. Yeah.
[36:02] (2162.24s)
Just just it's a new tool in our it's
[36:04] (2164.40s)
honestly a magical superpower because
[36:06] (2166.00s)
now it just unlocks so many things that
[36:07] (2167.44s)
you can do on top of it. Yeah. Yeah, but
[36:09] (2169.28s)
it feels a bit like, you know, the Harry
[36:10] (2170.72s)
Potter wand. Like when when you watch
[36:11] (2171.92s)
the movies, like, you know, at first it
[36:14] (2174.16s)
sounds magical when you read the book,
[36:15] (2175.44s)
like you can do all these spells, but if
[36:17] (2177.12s)
you're a hardcore Harry Potter fan, you
[36:19] (2179.12s)
will know that there's only certain
[36:20] (2180.40s)
spells that you can do. And you know,
[36:22] (2182.48s)
there's a certain thing that you need to
[36:23] (2183.76s)
say. And so there's a there's a whole
[36:25] (2185.12s)
mechanic around it. And like for for
[36:27] (2187.20s)
every fantasy book as well, when there's
[36:29] (2189.28s)
a magical world, like there are the
[36:30] (2190.56s)
rules and there's people who can master
[36:32] (2192.00s)
those rules. And I I feel it's a bit the
[36:34] (2194.24s)
same this, right? It's at first it's
[36:36] (2196.00s)
magic, but actually it has the rules and
[36:37] (2197.36s)
once you learn it, you can you can be
[36:38] (2198.96s)
this, you know, sorcerer who can
[36:42] (2202.00s)
Yeah, exactly. This episode is brought
[36:44] (2204.48s)
to you by
[36:45] (2205.40s)
cortex.io. Still tracking your services
[36:47] (2207.52s)
and production readiness in a
[36:48] (2208.72s)
spreadsheet. Real microservices named
[36:50] (2210.80s)
after TV show characters. You aren't
[36:53] (2213.12s)
alone. Being woken up at 3 a.m. for an
[36:55] (2215.60s)
incident and trying to figure out who
[36:56] (2216.88s)
owns what service, that's no fun. Cortex
[37:00] (2220.00s)
is the internal developer portal that
[37:01] (2221.52s)
solves service ownership and accelerates
[37:03] (2223.36s)
the path to engineering excellence
[37:05] (2225.36s)
within minutes. Determine who owns each
[37:07] (2227.36s)
service with Cortex's AI service
[37:09] (2229.04s)
ownership model even across thousands of
[37:11] (2231.96s)
repositories. Clear ownership means
[37:13] (2233.92s)
faster migrations, quicker resolutions
[37:15] (2235.84s)
to critical issues like lock forj and
[37:17] (2237.92s)
fewer adhere pings during incidents.
[37:20] (2240.96s)
Cortex is trusted by leading engineing
[37:22] (2242.88s)
organizations like affirm trip advisor,
[37:24] (2244.88s)
grammarly and sofi. Solve service
[37:27] (2247.36s)
ownership and unlock your team's full
[37:29] (2249.04s)
potential with Cortex. Visit
[37:31] (2251.96s)
cortex.io/pragmatic to learn more. That
[37:37] (2257.40s)
cotx.io/pragmatic. So then you had a
[37:39] (2259.36s)
really great run at uh KOD and then you
[37:43] (2263.04s)
did something like you decided to to to
[37:46] (2266.00s)
look look around the market and you
[37:47] (2267.92s)
blogged about this. You interviewed at
[37:50] (2270.44s)
46 companies. Did I get that right? Yes.
[37:54] (2274.00s)
But there is context behind that. I I'd
[37:56] (2276.56s)
love love love to understand like how
[37:58] (2278.16s)
you went about interviewing uh
[37:59] (2279.76s)
especially specifically for for an AI
[38:01] (2281.36s)
position. What did you learn about what
[38:03] (2283.60s)
the market is like, what interviewing is
[38:05] (2285.92s)
like, what the what what the whole whole
[38:08] (2288.00s)
scene is and if you can give a little
[38:09] (2289.28s)
context on like where you did this in
[38:10] (2290.88s)
terms of location wise types of
[38:12] (2292.40s)
companies just to help you know us us
[38:14] (2294.56s)
all understand this. Sure. Maybe just by
[38:17] (2297.12s)
giving a little bit of context. It was
[38:19] (2299.04s)
over a six-month period and the first
[38:21] (2301.60s)
half I wasn't closing them. I was
[38:23] (2303.76s)
getting noses. I was getting ramped up
[38:25] (2305.20s)
on my leak coded system design prep.
[38:27] (2307.04s)
After that, the interview process did
[38:29] (2309.28s)
take longer than I expected though
[38:31] (2311.28s)
because the AI space is especially noisy
[38:33] (2313.92s)
right now. And when I was trying to do
[38:35] (2315.60s)
my due diligence, like we were talking
[38:37] (2317.20s)
about earlier, there were often open
[38:40] (2320.00s)
questions that made me feel uneasy about
[38:42] (2322.00s)
the growth potential. And the advice I
[38:43] (2323.92s)
got from a mentor was that if it's not
[38:45] (2325.68s)
heck yes and if you have savings, don't
[38:47] (2327.84s)
join. It's not fair to the company or
[38:49] (2329.28s)
you. So that was how I thought about
[38:51] (2331.12s)
this. In terms of the space, it was
[38:56] (2336.00s)
clear that there are the product
[38:57] (2337.28s)
companies, infrastructure companies, and
[38:59] (2339.20s)
the model companies. I found it helpful
[39:01] (2341.12s)
to put companies in each category and
[39:05] (2345.20s)
figure out which segment you're most
[39:07] (2347.20s)
excited about to help narrow down the
[39:09] (2349.68s)
options given that there's so many AI
[39:11] (2351.28s)
companies right now. Could you give just
[39:12] (2352.80s)
an example of of each, especially with
[39:14] (2354.80s)
the infer model? I think it might be a
[39:16] (2356.80s)
bit I'm interested in how you're
[39:18] (2358.40s)
thinking about that. Yeah, product
[39:20] (2360.00s)
companies are the companies building on
[39:21] (2361.36s)
top of the model. Here I think of
[39:23] (2363.28s)
cursor, kodium, heia. Infrastructure
[39:26] (2366.00s)
companies are the companies building the
[39:28] (2368.00s)
tools to help AI product companies
[39:31] (2371.60s)
effectively use LLMs. So whole suite of
[39:34] (2374.56s)
these there are the inference providers
[39:36] (2376.08s)
like modal fireworks together vector
[39:38] (2378.72s)
database companies like pine cone
[39:40] (2380.56s)
chromodb weate evalid observability
[39:43] (2383.28s)
tools like brain trust arise galo and a
[39:46] (2386.16s)
whole other suite of products and then
[39:49] (2389.20s)
there's the model companies which are
[39:50] (2390.80s)
the base of the ecosystem building the
[39:52] (2392.56s)
intelligence you have the big tech
[39:54] (2394.56s)
companies like Google meta building
[39:57] (2397.12s)
models and then you also have startups
[39:58] (2398.80s)
like or not startups you have other big
[40:00] (2400.72s)
smaller companies open anthropic
[40:03] (2403.84s)
building models as well. I I I think
[40:05] (2405.60s)
it's a really good way to think about it
[40:07] (2407.60s)
and again I don't think many of us have
[40:09] (2409.60s)
uh verbalized it uh like this or this
[40:12] (2412.80s)
also goes back to to not many people
[40:15] (2415.04s)
have necessarily gone through I will say
[40:16] (2416.88s)
I this is not uh something that I came
[40:18] (2418.56s)
up with myself Yashkamar uh a mentor he
[40:20] (2420.96s)
pointed out that you should look at the
[40:22] (2422.16s)
space like this and that's how I think
[40:23] (2423.84s)
about it now. Wonderful. And what did
[40:25] (2425.60s)
you learn about like each of these
[40:27] (2427.28s)
companies in terms of the interview
[40:29] (2429.36s)
process, what the vibe was
[40:31] (2431.80s)
like like generically and also like how
[40:34] (2434.48s)
how you personally felt about it because
[40:36] (2436.08s)
like as I understand where you were uh
[40:38] (2438.32s)
Kota we can put them in the product
[40:40] (2440.16s)
category sorry the product company
[40:42] (2442.00s)
category. So for me in trying to be more
[40:45] (2445.12s)
focused in my search I decided to focus
[40:47] (2447.92s)
on model and infrastructure companies
[40:49] (2449.68s)
because I wanted to keep getting breath
[40:52] (2452.16s)
in what I was doing and I felt like the
[40:53] (2453.76s)
product companies were too similar to my
[40:55] (2455.44s)
experience at KOD which was phenomenal
[40:57] (2457.20s)
but I wanted to keep growing and that
[40:59] (2459.36s)
definitely the trade-off was that it's a
[41:01] (2461.20s)
bit more of an uphill battle because the
[41:02] (2462.72s)
work that I had done was not as relevant
[41:04] (2464.80s)
to model or infrastructure companies. In
[41:07] (2467.36s)
terms of the vibe, I think all of them
[41:09] (2469.36s)
are shipping really fast, have really
[41:11] (2471.84s)
lean teams, and are out to win. So, it's
[41:15] (2475.20s)
a very exciting time to be looking at
[41:16] (2476.72s)
them. Questions I would ask myself when
[41:18] (2478.88s)
I was trying to figure out, is this
[41:20] (2480.72s)
company viable in the long run on the
[41:23] (2483.76s)
infrastructure side was are their
[41:25] (2485.36s)
margins high enough given that so many
[41:27] (2487.04s)
of these infra inference
[41:29] (2489.48s)
providers are also paying for expensive
[41:32] (2492.40s)
GPUs. So, what is the margins here?
[41:34] (2494.56s)
especially when a good software business
[41:36] (2496.40s)
should have about 70% gross margins. And
[41:39] (2499.44s)
how easy is it to build infrastructure
[41:41] (2501.44s)
in house? You know, we know this, but
[41:43] (2503.28s)
engineers are a hard group of people to
[41:45] (2505.52s)
sell to because if it's not complex
[41:48] (2508.24s)
enough, if it's too expensive or doesn't
[41:50] (2510.40s)
work exactly how they want, engineers
[41:52] (2512.40s)
will just build it in-house. Google's a
[41:54] (2514.08s)
phenomenal example that's built so much
[41:56] (2516.00s)
inhouse. So, that's how I was thinking
[41:57] (2517.84s)
about the infrastructure companies. In
[42:00] (2520.08s)
terms of the model companies, I was just
[42:01] (2521.60s)
trying to get a sense of if they're
[42:03] (2523.04s)
training Frontier models, can they
[42:04] (2524.96s)
afford to keep training them given how
[42:07] (2527.20s)
expensive it is? Are they staying ahead
[42:09] (2529.60s)
of the open-source competition? Because
[42:12] (2532.24s)
if they're open weights that exist for a
[42:15] (2535.28s)
model, no one's going to want to pay a
[42:16] (2536.80s)
premium to get the model from a closed
[42:19] (2539.36s)
source provider. It's it's a sad
[42:21] (2541.20s)
reality. It is. And I think that it's
[42:23] (2543.60s)
interesting because today product
[42:25] (2545.60s)
companies are still willing to pay a
[42:27] (2547.52s)
premium for the best model even though
[42:29] (2549.68s)
an open weight exists as long as the the
[42:34] (2554.48s)
closed source provider is ahead. Yes.
[42:36] (2556.64s)
And anyone who's not nodding along when
[42:38] (2558.96s)
they'll find themselves evaluating an
[42:40] (2560.96s)
offer or a company and trying to
[42:42] (2562.80s)
understand the margins, that's a hard
[42:44] (2564.32s)
one to do, especially as an engineer.
[42:46] (2566.24s)
Where where did you get like data or did
[42:48] (2568.48s)
did companies answer some of your your
[42:50] (2570.16s)
questions on on the unit economics? Is
[42:52] (2572.24s)
these are things that companies like to
[42:54] (2574.00s)
have under wraps even as someone who's
[42:56] (2576.40s)
covering sometimes these these companies
[42:57] (2577.84s)
are just interested in the space even
[43:00] (2580.16s)
publications uh like financial
[43:02] (2582.08s)
publications you know will will just
[43:03] (2583.60s)
kind of wave their their hands because
[43:05] (2585.52s)
it is hard like this is the this is the
[43:07] (2587.44s)
big question and and these companies
[43:08] (2588.80s)
they want to hide these things from the
[43:11] (2591.12s)
casual observer for sure. Exactly. I
[43:14] (2594.16s)
think it's totally fair for a company
[43:15] (2595.52s)
not to share this information until you
[43:17] (2597.44s)
get an offer because it is sensitive
[43:19] (2599.04s)
information. I do think once you have an
[43:20] (2600.88s)
offer, it would be irresponsible for
[43:22] (2602.72s)
them not to tell you when you are as an
[43:24] (2604.48s)
investor as well and you sign an NDA. So
[43:26] (2606.24s)
you keep it to yourself. So I do think
[43:27] (2607.68s)
they should tell you for questions or
[43:30] (2610.16s)
for companies in the inference space. I
[43:32] (2612.08s)
would just ask you know how much money
[43:33] (2613.36s)
do you spend on the GPUs and then how
[43:35] (2615.36s)
much revenue do you have to make rough
[43:37] (2617.04s)
back of the envelope math of what those
[43:39] (2619.92s)
margins are like to just get some sense
[43:41] (2621.76s)
of the business. And then I also found
[43:44] (2624.64s)
it helpful to read some news providers
[43:46] (2626.88s)
like the information that does very good
[43:49] (2629.36s)
diligence on the economics behind
[43:52] (2632.88s)
different startups in the AI space. And
[43:55] (2635.36s)
if I could, I would try to also ask
[43:57] (2637.64s)
investors who have invested in these
[44:00] (2640.48s)
companies or passed on investing in
[44:02] (2642.72s)
these companies because they see the uh
[44:06] (2646.00s)
deck come to them. So they have a lot
[44:07] (2647.68s)
more insight onto what the business
[44:10] (2650.24s)
really looks like.
[44:11] (2651.92s)
you're talking like a like an investor
[44:14] (2654.24s)
or or or or like how a senior executive
[44:17] (2657.28s)
would do it which I love. I think more
[44:19] (2659.20s)
people should be doing this by the way
[44:21] (2661.04s)
and not enough people are are doing it.
[44:22] (2662.96s)
So it it's just very refreshing to hear
[44:25] (2665.60s)
and as and by the way like the investor
[44:28] (2668.32s)
is interesting because in my experience
[44:29] (2669.52s)
investors when you are applying to a
[44:31] (2671.76s)
company that they're investor in they
[44:32] (2672.96s)
actually want to help close great people
[44:34] (2674.64s)
and they will Exactly. they will happily
[44:37] (2677.12s)
connect and then you also have a
[44:38] (2678.72s)
connection where a few years down the
[44:40] (2680.64s)
road that investor might reach out to
[44:42] (2682.72s)
you saying oh I remember you're you
[44:44] (2684.40s)
you're a business-minded engineer but
[44:46] (2686.40s)
you know like in the future it's hard to
[44:47] (2687.92s)
tell I think we were talking about this
[44:49] (2689.28s)
before what will be in the future but
[44:50] (2690.96s)
there will be only more demand for
[44:52] (2692.72s)
software engineers who not only know how
[44:54] (2694.24s)
to code but are curious about the
[44:56] (2696.00s)
business can communicate with users etc
[44:58] (2698.56s)
so you'll now have a network a stronger
[45:00] (2700.80s)
network so there's only upside in in
[45:02] (2702.48s)
doing your due diligence it can actually
[45:04] (2704.00s)
help your career that's true And I 100%
[45:06] (2706.72s)
agree with investors being very generous
[45:08] (2708.64s)
with their time in wanting to chat with
[45:10] (2710.24s)
you and explain to you how the business
[45:11] (2711.92s)
works. So that's been something that's
[45:14] (2714.00s)
been really fun to do for sure. And then
[45:16] (2716.40s)
uh just going back to like this is all
[45:18] (2718.40s)
great when you get an offer, but how did
[45:19] (2719.84s)
you get to getting an offer? Like what
[45:21] (2721.20s)
what what did you need to brush up on in
[45:22] (2722.80s)
terms of interviews? Was it the pretty
[45:24] (2724.96s)
typical you know tech interviews even
[45:26] (2726.64s)
though these were for AI engineering
[45:28] (2728.08s)
roles of the the lead code system design
[45:30] (2730.40s)
or were there some AI specific things?
[45:32] (2732.08s)
you know what what what helped you go
[45:33] (2733.68s)
from initially you stumbled and you
[45:35] (2735.36s)
didn't get too much to like okay you
[45:37] (2737.12s)
actually like we're getting offers now
[45:39] (2739.36s)
in terms of the interview process I
[45:41] (2741.60s)
definitely thought it was all over the
[45:43] (2743.68s)
place as the market is trying to move
[45:48] (2748.00s)
away from leak code but still asks leak
[45:50] (2750.48s)
code so then you end up having to study
[45:52] (2752.32s)
leak code as well unless you know
[45:53] (2753.92s)
exactly where you're applying to so
[45:56] (2756.32s)
there were coding interviews system
[45:58] (2758.24s)
design and then projects coding was a of
[46:02] (2762.24s)
data structures and algorithms where the
[46:04] (2764.00s)
best way to do it is leak code. Luckily,
[46:06] (2766.96s)
neat code with an N now exists and he
[46:09] (2769.04s)
has phenomenal videos. So that was
[46:10] (2770.72s)
great. I believe in doing space
[46:12] (2772.24s)
repetition. So doing those questions a
[46:13] (2773.76s)
lot of times. Then there were front-end
[46:16] (2776.16s)
questions because I'm I'm full stack
[46:17] (2777.52s)
engineer as well. And I found that there
[46:19] (2779.44s)
was this resource, the great front end,
[46:21] (2781.12s)
that had lovely interview questions for
[46:23] (2783.68s)
the obscure JavaScript questions they
[46:25] (2785.60s)
sometimes ask.
[46:27] (2787.60s)
on the back end. That one I just more
[46:29] (2789.28s)
relied on things that I had done at work
[46:31] (2791.20s)
for those interviews. That's the coding
[46:33] (2793.12s)
part. The system design part, I thought
[46:35] (2795.20s)
Alex Shu, system design, his two books,
[46:37] (2797.20s)
phenomenal. Just reading those, really
[46:39] (2799.28s)
understanding them, doing them again and
[46:40] (2800.64s)
again until you understand why things
[46:43] (2803.76s)
work a certain way. Honestly, I love
[46:45] (2805.36s)
system design interviews. They're really
[46:47] (2807.12s)
fun because you learn things outside of
[46:48] (2808.88s)
the the domain that you're in as well.
[46:50] (2810.40s)
And then there are the third type of
[46:51] (2811.68s)
interviews, which is project interviews
[46:53] (2813.28s)
where go build something in a day. And
[46:55] (2815.28s)
those are probably my favorite out of
[46:57] (2817.60s)
all of them because you get to show how
[46:58] (2818.96s)
passionate you are about that specific
[47:00] (2820.88s)
product and you can actually show what
[47:02] (2822.64s)
you can do. I do hope that as an
[47:04] (2824.80s)
industry we move away from leak code and
[47:08] (2828.00s)
instead move to just project interviews
[47:09] (2829.84s)
reading code which has become way more
[47:11] (2831.44s)
important today as well as debugging
[47:12] (2832.96s)
code. But I think we're kind of in the
[47:14] (2834.64s)
interim where as an industry we haven't
[47:17] (2837.12s)
fully formed an opinion here. And then
[47:19] (2839.28s)
most of these interviews was it the end
[47:20] (2840.64s)
of the last year? So end of 2024 or so
[47:24] (2844.24s)
were they were they remote or or were
[47:26] (2846.08s)
some some were in person already? Swiss
[47:29] (2849.40s)
in between June of last year and a large
[47:34] (2854.00s)
chunk were remote but there were
[47:35] (2855.60s)
definitely interviews in person as well
[47:38] (2858.32s)
which I enjoy because I was very much
[47:40] (2860.56s)
optimizing for companies that are in
[47:42] (2862.48s)
person. Yeah, we we we'll see. But I
[47:45] (2865.04s)
think we're sensing a trend or I'm
[47:46] (2866.32s)
sensing a trend that inerson injuries
[47:48] (2868.08s)
might be starting to go back at least
[47:49] (2869.12s)
your final rounds which by the way it
[47:51] (2871.28s)
might not be a bad I mean it's
[47:52] (2872.64s)
interesting because before co like uh
[47:54] (2874.56s)
like when you know I spent like most of
[47:57] (2877.28s)
my career there it was just in person
[47:59] (2879.28s)
and there are so many upsides right you
[48:01] (2881.28s)
do meet the people you do see the
[48:02] (2882.88s)
location often times you meet your
[48:04] (2884.48s)
future teammates and for example for me
[48:07] (2887.28s)
I I once in London I had uh two offers
[48:10] (2890.32s)
between two two banks and in one case I
[48:13] (2893.84s)
met my future team the whole team and
[48:16] (2896.24s)
when I didn't meet my future team it was
[48:17] (2897.84s)
just like they said like you will be
[48:19] (2899.04s)
assigned a team and I actually chose it
[48:21] (2901.28s)
was a lower salary, but I chose a lower
[48:23] (2903.20s)
salary cuz I really like the people. And
[48:25] (2905.20s)
you know, like we just kicked it off. It
[48:26] (2906.64s)
it felt like a good connection. And back
[48:28] (2908.88s)
then I went through a recruiter, so the
[48:30] (2910.56s)
recruiter negotiated the same salary for
[48:32] (2912.16s)
me, which was kind of a win, I I guess.
[48:35] (2915.20s)
But like there there are like I I know
[48:37] (2917.28s)
there's you know like it's it's always
[48:39] (2919.60s)
we will hear people like mourning the
[48:41] (2921.52s)
the end of or or fewer remote interviews
[48:44] (2924.08s)
but there are all these upsides which
[48:46] (2926.08s)
when you're committing to a place for so
[48:47] (2927.60s)
many for hopefully many years you want
[48:50] (2930.32s)
to have all that information 100%.
[48:52] (2932.56s)
Definitely I think it's energizing on
[48:54] (2934.32s)
both ends for sure. It's a great point.
[48:56] (2936.08s)
And so in the end you joined open AI
[48:58] (2938.80s)
right? Yes I did. Congratulations. Thank
[49:01] (2941.92s)
you. And then can you share on what kind
[49:04] (2944.32s)
of general work you do at OpenAI? Sure.
[49:07] (2947.20s)
So I work on safety as an engineer at
[49:09] (2949.68s)
OpenAI and OpenAI's goal and mission is
[49:12] (2952.40s)
to build AGI that benefits all of
[49:14] (2954.24s)
humanity. On safety we focus on the
[49:16] (2956.40s)
suffix of that statement. So benefiting
[49:18] (2958.80s)
all of humanity. Some things I work on
[49:22] (2962.68s)
are a small low latency classifiers that
[49:28] (2968.32s)
detect when the model or users are doing
[49:31] (2971.28s)
things that are harmful so that you can
[49:32] (2972.64s)
block live. So that means the training
[49:35] (2975.68s)
the data flywheel hosting these models
[49:37] (2977.76s)
to scale. Second thing that I get to
[49:40] (2980.56s)
work on is measuring when the models are
[49:43] (2983.60s)
being harmful in the wild. And there are
[49:45] (2985.52s)
a lot of dual use cases over here. But
[49:47] (2987.92s)
really trying to get a sense as these
[49:49] (2989.52s)
models become more capable and people
[49:51] (2991.36s)
are figuring out different ways to
[49:52] (2992.88s)
jailbreak them and exploit them. What
[49:55] (2995.28s)
are those unknown harms that we don't
[49:57] (2997.76s)
know of with more powerful models and
[50:00] (3000.72s)
then distilling it into small
[50:02] (3002.48s)
classifiers. There's also on my team a
[50:04] (3004.64s)
lot of safety mitigation services that
[50:06] (3006.64s)
we own. And so part of our work is to
[50:09] (3009.36s)
integrate it with all the different
[50:10] (3010.96s)
product launches and as you know there
[50:12] (3012.64s)
are a lot of different product launches
[50:13] (3013.84s)
that definitely keeps our team busy and
[50:15] (3015.76s)
that's just the tip of the surface.
[50:16] (3016.88s)
There are a lot more things that we work
[50:17] (3017.84s)
on at safety. I mean this is like it
[50:20] (3020.40s)
sounds very interesting because when I
[50:21] (3021.84s)
worked on payments back at at Uber we
[50:24] (3024.00s)
had a team called fraud and oh boy they
[50:26] (3026.40s)
had so many stories like I just talking
[50:28] (3028.48s)
with them I like you would think you
[50:31] (3031.20s)
know like payments is pretty simple like
[50:33] (3033.12s)
oh you just need to pay but then the
[50:34] (3034.48s)
edge cases are always interesting with
[50:36] (3036.52s)
every every area and the same thing I
[50:39] (3039.12s)
guess with I mean LM are not as simple
[50:41] (3041.28s)
but once you realize how they work next
[50:42] (3042.80s)
token prediction it sounds pretty simple
[50:44] (3044.24s)
but then the edge cases and all the
[50:45] (3045.60s)
things that could go wrong etc. And it
[50:46] (3046.88s)
sounds like you're kind of in the middle
[50:48] (3048.08s)
of of that like having a very like good
[50:50] (3050.24s)
vantage point in actually in in in in
[50:52] (3052.32s)
the details. You've now worked at KOD.
[50:54] (3054.48s)
You've you've interned at at Google and
[50:56] (3056.00s)
and Microsoft and you talk with mentors
[50:58] (3058.00s)
about like what other places are. What
[51:00] (3060.00s)
are things that you feel that are just
[51:01] (3061.44s)
very kind of distinctly different about
[51:03] (3063.28s)
OpenAI compared to other companies? I
[51:06] (3066.08s)
think what makes OpenAI unique is the
[51:09] (3069.04s)
mix of speed and building for scale. You
[51:12] (3072.56s)
know at startups you get that speed of
[51:14] (3074.40s)
iteration and it's so fun and then at
[51:16] (3076.96s)
bigger places you get to build for scale
[51:19] (3079.44s)
but OpenAI is in a very unique spot
[51:21] (3081.20s)
where you have both at the moment things
[51:22] (3082.96s)
move really fast and you have huge
[51:26] (3086.48s)
amounts of users. The service that I
[51:29] (3089.36s)
work on you know gets 60k requests per
[51:31] (3091.44s)
second and you just think normally you
[51:34] (3094.40s)
get one or the other and it's really fun
[51:36] (3096.88s)
to get both.
[51:39] (3099.44s)
Second thing that I think is quite
[51:41] (3101.68s)
unique for a company of this size is the
[51:45] (3105.04s)
open culture. People are very open to
[51:46] (3106.64s)
answering questions on why and how
[51:49] (3109.60s)
things work a certain way. So, it's a
[51:51] (3111.84s)
great place to learn, which is something
[51:54] (3114.56s)
I didn't realize from the outside. And
[51:56] (3116.32s)
then third, people are just very
[51:58] (3118.96s)
passionate about the mission, work
[52:00] (3120.08s)
really hard. And I don't think this is
[52:01] (3121.28s)
unique to OpenAI in and of itself. All
[52:03] (3123.68s)
companies, I think, where great work is
[52:05] (3125.12s)
happening, people are like this. But
[52:07] (3127.12s)
it's just never a boring day in the
[52:09] (3129.52s)
office because people care so much and
[52:11] (3131.60s)
are constantly shipping. Yeah. And then
[52:14] (3134.24s)
talk about shipping. Like you've I I'm
[52:16] (3136.88s)
assuming you you ship some things to
[52:18] (3138.48s)
production already, but how can we
[52:19] (3139.92s)
imagine a thing a project an idea making
[52:23] (3143.12s)
into production, right? Like there's a
[52:24] (3144.96s)
there's a very bureaucratic companies,
[52:27] (3147.20s)
you know, I don't want to like say old
[52:28] (3148.88s)
Microsoft, maybe maybe not today, but
[52:30] (3150.80s)
where there's like, you know, like very
[52:32] (3152.72s)
strict planning process. Then Jer
[52:34] (3154.88s)
tickets are created by the PM. The
[52:36] (3156.56s)
engineers have to pick it up. Then
[52:38] (3158.64s)
someone else might actually deploy it.
[52:40] (3160.16s)
So like this is the super like old
[52:42] (3162.56s)
school and slow and and the reason why
[52:44] (3164.56s)
some engineers don't like it. What is it
[52:46] (3166.72s)
like? You mentioned it's fast, but what
[52:49] (3169.20s)
was your experience in getting things
[52:51] (3171.44s)
from idea to production? And is it
[52:53] (3173.36s)
multiple teams? Can one person actually
[52:55] (3175.20s)
do it? Is it even allowed? I I I don't
[52:57] (3177.28s)
know. I think it's very much allowed and
[52:59] (3179.20s)
very much encouraged.
[53:01] (3181.52s)
There's been publications of how deep
[53:05] (3185.60s)
research came to be where it was an
[53:07] (3187.60s)
engineer hacking on something presenting
[53:09] (3189.28s)
it to larger sea suite and now becoming
[53:11] (3191.52s)
a full very impactful product. So it's
[53:13] (3193.52s)
definitely encouraged which I love. I
[53:15] (3195.92s)
too have had a similar experience and
[53:17] (3197.52s)
it's very encouraged to come with ideas
[53:20] (3200.32s)
and actually drive them forward. just
[53:22] (3202.56s)
strictly from from your perspective,
[53:24] (3204.96s)
what what do you think like one thing
[53:26] (3206.64s)
that stands out that open AI can
[53:28] (3208.80s)
actually still ship so fast? Cuz it
[53:31] (3211.12s)
feels it defies a little bit the laws of
[53:33] (3213.20s)
the growing organization which
[53:34] (3214.72s)
eventually slows down. At one point, I'm
[53:36] (3216.80s)
sure it will, but there's no signs of
[53:38] (3218.40s)
this happening so far. My observation is
[53:41] (3221.28s)
that the systems are built to enable you
[53:44] (3224.84s)
to ship fast and they give engineers a
[53:47] (3227.68s)
lot of trust even though it comes with
[53:49] (3229.28s)
the downside of sometimes that can lead
[53:52] (3232.00s)
to outages. To put this very
[53:55] (3235.16s)
concretely, when I joined, you could
[53:57] (3237.44s)
make stats safe changes without an
[53:59] (3239.12s)
approval. So you have trust to go in and
[54:01] (3241.52s)
flip a flag to turn something on. That's
[54:03] (3243.60s)
no longer the case. You need one
[54:04] (3244.64s)
reviewer. The service that I get to work
[54:06] (3246.32s)
on has 60,000 requests per second, but
[54:09] (3249.76s)
you get to deploy with one review
[54:11] (3251.76s)
immediately. So my observation is that
[54:14] (3254.40s)
there is truly trust put in engineers to
[54:18] (3258.52s)
quickly and not have a lot of red tape
[54:20] (3260.96s)
around shipping fast. Yeah. And I think
[54:23] (3263.36s)
this just goes with uh kind of unset
[54:25] (3265.44s)
expectation that expectations will be
[54:27] (3267.04s)
very high of the people who come in here
[54:28] (3268.80s)
because you cannot hire an engineer who
[54:30] (3270.96s)
is used to you know being only doing a
[54:34] (3274.24s)
small part not used to thinking about
[54:35] (3275.92s)
the product and the business impact and
[54:37] (3277.60s)
all those things. So I have a sense that
[54:40] (3280.00s)
what you're doing it might be a kind of
[54:41] (3281.44s)
a given for you but in the industry it
[54:42] (3282.88s)
might be more common to expect that
[54:44] (3284.24s)
engineers are just wearing you know we
[54:45] (3285.84s)
used to call it wearing more hats but
[54:47] (3287.28s)
it's just like it's just how it is like
[54:49] (3289.20s)
you you you do want to have a you know
[54:52] (3292.24s)
like you're kind of merging a little bit
[54:53] (3293.84s)
of PM a data scientist an engineer allin
[54:56] (3296.56s)
one and these are the type of people who
[54:58] (3298.64s)
can actually make something like like
[55:01] (3301.52s)
open AI or or similar companies like
[55:03] (3303.28s)
work so well with with this many people.
[55:05] (3305.52s)
Yeah. And I just think with
[55:07] (3307.96s)
intelligence today, the roles between
[55:10] (3310.96s)
data science, engineering, backend,
[55:13] (3313.92s)
front end, PM blurs so much that each
[55:17] (3317.24s)
individual, whether you're an open air
[55:19] (3319.20s)
or not, is expected to do more of that
[55:21] (3321.36s)
because you can get help from a very
[55:24] (3324.56s)
capable model. And I think that makes it
[55:26] (3326.24s)
very exciting for us because it means
[55:27] (3327.52s)
that we can truly be full stack
[55:28] (3328.88s)
engineers and go from an idea to launch
[55:32] (3332.00s)
very quickly. Absolutely. So what what
[55:35] (3335.12s)
are some things that you've learned
[55:36] (3336.16s)
about uh AI engineering the kind of the
[55:38] (3338.56s)
realities of it because it's a very new
[55:40] (3340.48s)
field and like what are some surprising
[55:42] (3342.16s)
things that you you didn't quite expect?
[55:44] (3344.16s)
One thing that I've learned that I
[55:46] (3346.96s)
didn't realize coming in was how much of
[55:50] (3350.24s)
AI engineering is about building
[55:52] (3352.80s)
solutions to known limitations of the
[55:55] (3355.52s)
model and then as the model gets better
[55:57] (3357.36s)
you scrap that work and build new
[55:59] (3359.84s)
guardrails. Let me give an example from
[56:01] (3361.84s)
KOD. So prefunction calling days, we
[56:05] (3365.12s)
wanted a way to get our model to take
[56:07] (3367.44s)
action based on what the user said.
[56:09] (3369.92s)
Function calling didn't exist. So we
[56:11] (3371.36s)
prompted the model to return JSON, parse
[56:13] (3373.72s)
that and actually deterministically call
[56:17] (3377.28s)
an action based on that JSON blob. Then
[56:20] (3380.24s)
OpenAI released function calling. Okay,
[56:22] (3382.40s)
scrap that and instead integrate with
[56:24] (3384.00s)
function calling. But you know back in
[56:26] (3386.00s)
those days, function calling was not
[56:27] (3387.20s)
very reliable. And now today we moved
[56:29] (3389.60s)
from function calling to the MCP
[56:31] (3391.96s)
paradigm. So things are changing very
[56:35] (3395.76s)
quickly and the models are getting
[56:37] (3397.60s)
better but they're still not perfect.
[56:39] (3399.28s)
The moment you get more capability,
[56:41] (3401.20s)
there are more engineering guardrails
[56:42] (3402.80s)
you need to build to make sure they work
[56:44] (3404.40s)
reliably at scale. Yeah. And I guess you
[56:47] (3407.04s)
need to become comfortable with throwing
[56:48] (3408.96s)
away your work when the model is there.
[56:52] (3412.00s)
you just need to not be as attached to
[56:53] (3413.60s)
it cuz I I think there's a little bit of
[56:55] (3415.04s)
this especially when you're used to like
[56:56] (3416.64s)
things not changing as much as software
[56:58] (3418.08s)
engineering. So just you know like it's
[57:00] (3420.32s)
it's it's not a it's not a waste it's a
[57:02] (3422.56s)
learning. Yeah. And it's just been
[57:05] (3425.28s)
easier now to or cheaper to produce code
[57:08] (3428.80s)
and so you see this expansion and
[57:11] (3431.20s)
collapse phase happen a lot and where
[57:13] (3433.36s)
you build a lot of features see what
[57:14] (3434.88s)
works and then collapse to what works
[57:16] (3436.32s)
and restart. There's a lot of scrapping
[57:18] (3438.72s)
your work as the creation of code
[57:20] (3440.64s)
becomes cheaper. It's easier not to be
[57:22] (3442.64s)
attached when an LLM also helped
[57:24] (3444.48s)
generate that code. Yeah, I I think this
[57:26] (3446.40s)
will be a big change, a good change once
[57:29] (3449.60s)
we get used to it. Yes, exactly. Now,
[57:32] (3452.48s)
when it comes to AI and junior
[57:33] (3453.84s)
engineers, uh like you're such an
[57:36] (3456.16s)
interesting example in the sense that
[57:37] (3457.60s)
you you started your career a little bit
[57:39] (3459.28s)
before AI took off. Uh but you you also
[57:42] (3462.72s)
transitioned with like not not decades
[57:46] (3466.08s)
of of experience just yet. What what is
[57:48] (3468.56s)
your take on how Gen AI will impact new
[57:52] (3472.68s)
grads, people who are still in college?
[57:54] (3474.80s)
Cuz you know there there's two takes and
[57:56] (3476.24s)
they're both very extreme. One is the
[57:58] (3478.80s)
engineers with 10 plus years experience
[58:00] (3480.48s)
often just feel like I feel so sorry for
[58:02] (3482.24s)
these people like they're they're not
[58:03] (3483.92s)
going to get jobs even if they get jobs.
[58:05] (3485.68s)
They're not going to dependent on AI.
[58:07] (3487.12s)
They're not going to like read the
[58:08] (3488.16s)
books. you won't know what it was exact
[58:10] (3490.24s)
back in our day, right? So, so there
[58:12] (3492.00s)
there's this this thing and and also
[58:15] (3495.28s)
like some some people are generally
[58:17] (3497.76s)
worried that well, you know, you can now
[58:19] (3499.28s)
outsource so outsource so many things to
[58:21] (3501.04s)
AI. They're thinking, okay, maybe they
[58:22] (3502.40s)
can pick up things really quickly, but
[58:24] (3504.16s)
maybe they're not never going to get to
[58:25] (3505.44s)
that depth. Now, I think that both are
[58:27] (3507.20s)
extreme. I I'd love to hear like how you
[58:29] (3509.28s)
see it because you're kind of seeing
[58:30] (3510.40s)
this firsthand. Definitely. And you're
[58:33] (3513.12s)
right, from my experience level, I get
[58:35] (3515.20s)
insight into what both of those
[58:37] (3517.24s)
engineering lives are like. And
[58:40] (3520.24s)
currently, I'm not convinced that AI is
[58:43] (3523.04s)
going to be disproportionately worse for
[58:45] (3525.20s)
junior engineers. In fact, I think that
[58:48] (3528.08s)
it allows everyone to move higher into
[58:50] (3530.72s)
the stack and be more creative in what
[58:54] (3534.72s)
you're building. You empower younger
[58:57] (3537.20s)
engineers to just do more, propose
[59:00] (3540.32s)
ideas, and actually ship that
[59:03] (3543.44s)
I do subscribe to the take that there
[59:05] (3545.52s)
will be people that use AI to learn and
[59:08] (3548.40s)
then people that use AI to avoid
[59:10] (3550.60s)
learning. I would say that there's
[59:12] (3552.64s)
actually room for both things to exist
[59:14] (3554.64s)
and you should be doing both. I
[59:16] (3556.56s)
personally think that when you're
[59:17] (3557.84s)
working on a green field project trying
[59:19] (3559.36s)
to prove a vision that something should
[59:21] (3561.56s)
exist, why not skip the learning vibe
[59:24] (3564.40s)
code it to actually get a real product
[59:26] (3566.72s)
that you can validate and then go run to
[59:30] (3570.16s)
build this for real as a new product
[59:31] (3571.68s)
line? But I don't think you should skip
[59:34] (3574.32s)
the learning when you're trying to now
[59:36] (3576.80s)
build a robust system that you are the
[59:38] (3578.96s)
owner of because when hits the fan
[59:41] (3581.84s)
and you're in a sev, AI doesn't help
[59:44] (3584.16s)
that much because it doesn't work very
[59:45] (3585.76s)
well in between at a high systems level
[59:48] (3588.32s)
and then you know reading logs. So when
[59:50] (3590.08s)
you own the code, I think you should use
[59:51] (3591.52s)
AI to learn to understand all the edge
[59:53] (3593.92s)
cases of why things work a certain way.
[59:56] (3596.56s)
So I think there's room for both. It's
[59:58] (3598.40s)
going to be an important skill for us to
[60:00] (3600.24s)
learn when we should
[60:01] (3601.96s)
outsource the doing versus when we
[60:05] (3605.76s)
should use AI to make ourselves better
[60:07] (3607.68s)
and stronger engineers. Yeah. And I
[60:09] (3609.68s)
guess like there's probably not too much
[60:11] (3611.44s)
harm in if you don't understand it,
[60:13] (3613.84s)
spend some time to understand it and AI
[60:15] (3615.92s)
will help you typically do this faster.
[60:19] (3619.04s)
So like I I I'm not sure if this is uh
[60:22] (3622.80s)
to do with personality or curiosity but
[60:26] (3626.64s)
we've seen this before by the way like
[60:28] (3628.32s)
when uh any time but let's say 10 10
[60:30] (3630.80s)
years before when I was like maybe a mid
[60:32] (3632.80s)
mid-level engineer like I saw new grads
[60:35] (3635.28s)
join the workforce and we were now using
[60:37] (3637.36s)
you know higher level languages like
[60:38] (3638.56s)
JavaScript or Python or or TypeScript or
[60:41] (3641.68s)
take take the example of of the the
[60:43] (3643.68s)
recent uh you know a few years ago like
[60:45] (3645.44s)
like new grad engineers they start to
[60:47] (3647.68s)
react And when you start with React, you
[60:50] (3650.08s)
JavaScript and Typescript. A lot of
[60:52] (3652.08s)
people who haven't studied computer
[60:54] (3654.80s)
science and didn't do assembly or or C++
[60:56] (3656.80s)
or these kind of things, you can just
[60:58] (3658.00s)
learn React and you can just stay there
[60:59] (3659.36s)
and you can figure out how to use it.
[61:01] (3661.44s)
But the be the better developers have
[61:03] (3663.84s)
always asked why does it work like this?
[61:06] (3666.72s)
What happens? What is a virtual DOM? How
[61:09] (3669.20s)
can I do it? How can I manipulate it?
[61:11] (3671.04s)
and and you look at the source code and
[61:13] (3673.52s)
I feel there's always been the people
[61:15] (3675.12s)
who do this and they're just better
[61:16] (3676.72s)
engineers eventually they can debug
[61:18] (3678.40s)
faster they they they ask why and you
[61:20] (3680.96s)
know they're slow so so I think in this
[61:22] (3682.56s)
new world we will just have this and I
[61:24] (3684.00s)
don't I don't think this trait will will
[61:25] (3685.36s)
die out in fact you know like I to me
[61:26] (3686.96s)
you're a great proof of you know you you
[61:29] (3689.20s)
go deep you understand how how the
[61:30] (3690.64s)
things work and then you figure you
[61:32] (3692.08s)
decide like okay I'm I'm I'm going to
[61:33] (3693.84s)
use it to my advantage right now I just
[61:35] (3695.68s)
want to go fast because I know what I'm
[61:37] (3697.20s)
doing already yes I do think that's spot
[61:39] (3699.36s)
on and that we've had this in in the
[61:40] (3700.88s)
past it will become even easier to
[61:44] (3704.24s)
outsource that intelligence and in some
[61:47] (3707.12s)
sense be lazy. So I think we'll have to
[61:49] (3709.20s)
just be more intentional as engineers to
[61:51] (3711.28s)
make sure that we are actually going
[61:53] (3713.88s)
deep in cases where it really matters.
[61:57] (3717.12s)
And so far from from what what you're
[61:59] (3719.20s)
seeing how because you've seen before
[62:01] (3721.92s)
before geni tools you're now working at
[62:04] (3724.16s)
a you've been a product engineer with
[62:06] (3726.32s)
with AI you're now working at a model
[62:07] (3727.92s)
company. How do you think these tools
[62:10] (3730.40s)
will change the software engine that you
[62:12] (3732.24s)
have been doing before? And how is it
[62:13] (3733.68s)
already changing your day-to-day work?
[62:16] (3736.00s)
In terms of what doesn't change, we
[62:17] (3737.84s)
still have code as the way innovation
[62:21] (3741.20s)
manifests. You go from idea to code to
[62:23] (3743.36s)
iterate. That's the same. As engineers,
[62:26] (3746.08s)
you still need to know how highle
[62:27] (3747.76s)
systems work and design them very well.
[62:31] (3751.12s)
You have to debug code really well and
[62:33] (3753.44s)
you have to be able to be really good at
[62:35] (3755.04s)
reading code. So to me that all stays
[62:37] (3757.60s)
the same. What's changed I think is the
[62:40] (3760.32s)
division of responsibilities between PM
[62:42] (3762.96s)
designer software engineer. I was
[62:44] (3764.72s)
talking to a friend at Decagon. I think
[62:46] (3766.16s)
they were telling me there are 100
[62:47] (3767.28s)
people and they still don't have a
[62:48] (3768.40s)
designer because product is just
[62:50] (3770.00s)
expected to do the design as well. As a
[62:53] (3773.76s)
software engineer, this has always been
[62:55] (3775.20s)
true at startups, but now more than
[62:56] (3776.96s)
ever, you're expected to do product work
[62:59] (3779.20s)
as well. We talked about this earlier.
[63:00] (3780.88s)
What also changes that software
[63:02] (3782.24s)
engineers become more full stack. You
[63:05] (3785.08s)
outsource work to another adjacent role
[63:08] (3788.48s)
like data engineer. You're expected to
[63:10] (3790.16s)
build those data pipelines yourself. I
[63:12] (3792.64s)
also think what's changed is that we
[63:14] (3794.24s)
need to be better at articulating our
[63:16] (3796.40s)
software engineering architectures and
[63:19] (3799.04s)
thoughts because you are expected to
[63:21] (3801.12s)
prompt models to do this. And the
[63:23] (3803.44s)
engineers that will be most efficient
[63:25] (3805.28s)
are the ones that can see the big
[63:28] (3808.92s)
picture. Write a great prompt that also
[63:31] (3811.68s)
catches the edge cases and then have the
[63:33] (3813.92s)
model implement it. It's like the best
[63:35] (3815.76s)
engineering managers that are able to
[63:37] (3817.12s)
zoom in and zoom out really well and
[63:40] (3820.16s)
being able to zoom out prompt what you
[63:42] (3822.08s)
need to do, but then zoom in when
[63:43] (3823.76s)
actually reading that code and catch
[63:45] (3825.20s)
potential bugs instead of just relying
[63:47] (3827.84s)
on the LLM to be 100% right in all cases
[63:51] (3831.20s)
because there'll be unique edge cases to
[63:53] (3833.36s)
the system that you're building that the
[63:54] (3834.64s)
LLM is not aware of and you need to be
[63:56] (3836.40s)
able to catch that when you're reading
[63:57] (3837.60s)
the code. Yeah, I I feel like if you
[64:00] (3840.72s)
have a mental model and I I see this so
[64:02] (3842.72s)
much when I'm using these tools, you
[64:04] (3844.16s)
know, when I'm either like vibe coding
[64:06] (3846.40s)
or or prompting or when I know what I
[64:08] (3848.64s)
want to do, when it's in my head, I I I
[64:11] (3851.12s)
either like because I know my code base
[64:13] (3853.44s)
or I know what I want to do or I just
[64:14] (3854.96s)
sat down, I I I thought through it and I
[64:17] (3857.12s)
drew it out. I'm I'm so fast. I'm great.
[64:19] (3859.28s)
Like and I can switch between, you know,
[64:20] (3860.96s)
I might do an agentic mode to like
[64:22] (3862.56s)
generate it. Maybe I like it, maybe I
[64:24] (3864.32s)
don't. Then I just do it by hand. It
[64:25] (3865.52s)
doesn't matter. Like I get there. like I
[64:27] (3867.04s)
know where I'm going, but when I don't I
[64:29] (3869.76s)
I did this where like oh I tried to vibe
[64:31] (3871.44s)
code a game and I failed not because I
[64:33] (3873.76s)
don't I just didn't know what I wanted
[64:34] (3874.96s)
to do like Yeah. And and you know when
[64:37] (3877.36s)
when your prompt doesn't tell like oh do
[64:39] (3879.52s)
this then I don't give it
[64:41] (3881.64s)
guidance. Yeah. Like it Yeah. No
[64:44] (3884.32s)
definitely it wasn't the fault of of the
[64:46] (3886.48s)
tool. It was just you know what what I
[64:48] (3888.72s)
don't know what I expect like how would
[64:51] (3891.28s)
this thing know which it's it's
[64:53] (3893.04s)
nondeterministic but you need to give us
[64:54] (3894.56s)
some direction. Exactly. For sure. And I
[64:57] (3897.52s)
also on that point think that it's great
[65:00] (3900.00s)
when you're zeroing and doing green
[65:01] (3901.92s)
field work, but today and I think this
[65:04] (3904.32s)
will change. It's not the best at
[65:06] (3906.16s)
working at in large code bases. And as
[65:08] (3908.24s)
engineers, you're always working in
[65:10] (3910.24s)
large code bases when you're building
[65:12] (3912.00s)
things for prod. And so that part of our
[65:14] (3914.72s)
job hasn't changed of being able to find
[65:18] (3918.16s)
the right place the code should go, use
[65:22] (3922.48s)
the right modules that exist and piece
[65:25] (3925.36s)
it together in a larger codebase when
[65:27] (3927.44s)
you're adding a feature. Yeah. Yeah. And
[65:28] (3928.72s)
also just like simple stuff which we
[65:30] (3930.00s)
take for granted but like setting up the
[65:31] (3931.60s)
the tools to like run the tests to to
[65:34] (3934.16s)
know how to deploy to know how to
[65:35] (3935.92s)
control the feature flags to how safely
[65:37] (3937.92s)
put out something so it doesn't go to
[65:39] (3939.60s)
prod if you want to AB test like which
[65:41] (3941.76s)
you know when you on board this is is a
[65:43] (3943.28s)
pretty kind of given but if you work at
[65:45] (3945.04s)
a place that has like microservices or
[65:46] (3946.56s)
whatever like it's it's all and I I feel
[65:48] (3948.56s)
like there there's so so many other
[65:50] (3950.32s)
other things but I I love I love how you
[65:53] (3953.12s)
you summarize what will not change
[65:56] (3956.24s)
because I think that that is really
[65:57] (3957.76s)
important. And I love how how you
[65:59] (3959.36s)
brought up software architecture. I I've
[66:00] (3960.96s)
been thinking about like this recently.
[66:02] (3962.40s)
In fact, I've started to read some like
[66:04] (3964.24s)
really old software architecture books
[66:06] (3966.24s)
because there are some ideas that I'm
[66:08] (3968.72s)
not sure will change. I I I want to say
[66:10] (3970.64s)
this theory, but it might not change as
[66:12] (3972.48s)
much. What books are software
[66:13] (3973.68s)
architecture books are you reading at
[66:14] (3974.88s)
the moment? I I' I've been going through
[66:16] (3976.40s)
the middle man month. I've almost
[66:18] (3978.08s)
finished this. This is a real old one.
[66:20] (3980.80s)
And then I have so this is from the 90s.
[66:23] (3983.52s)
It's called software architecture and
[66:25] (3985.76s)
it's by Mary Shaw and David Garlon and
[66:30] (3990.68s)
Grady Grady Bouch who's a legend in
[66:33] (3993.44s)
software engineering and I interviewed
[66:35] (3995.04s)
him. He said that he thinks this is the
[66:36] (3996.80s)
single best software book. Now it's it's
[66:39] (3999.36s)
very thin and it's it's it's it's I
[66:42] (4002.56s)
think from 1995 or so. So I I I've just
[66:45] (4005.12s)
heard to read it but I'm interested in
[66:47] (4007.92s)
what 1996 which is 30 years ago. I'm
[66:50] (4010.96s)
just interested in what are the things
[66:52] (4012.64s)
that might have not changed. Clearly,
[66:54] (4014.08s)
some things will be dated, right? Like
[66:55] (4015.84s)
they're talking about Cororba, which is
[66:57] (4017.20s)
like this old Java framework that we
[66:59] (4019.28s)
don't use anymore, but some of the other
[67:01] (4021.04s)
things there's a lot of uh reflection
[67:03] (4023.52s)
with civil engineering. And this book
[67:04] (4024.80s)
was written when there was no real
[67:06] (4026.56s)
software architecture. So, they tried to
[67:08] (4028.16s)
define it, which I'm kind of thinking
[67:11] (4031.12s)
there might be some interesting ideas.
[67:12] (4032.72s)
So, I'm I'm I'm interested like on on
[67:15] (4035.68s)
what has not changed for the most part.
[67:17] (4037.60s)
Yes. No, that's a very nice approach of
[67:20] (4040.96s)
actually looking at the history to see
[67:22] (4042.56s)
what's not changed from then to now to
[67:24] (4044.40s)
then extend that line to what won't
[67:25] (4045.92s)
change in the future. I'd be very
[67:26] (4046.96s)
curious to see what you learn from that.
[67:28] (4048.16s)
Well, and also reinventing, right?
[67:29] (4049.52s)
Because I feel we will have to reinvent
[67:31] (4051.44s)
some parts of the stack and it's I think
[67:32] (4052.96s)
it's important to understand also I feel
[67:35] (4055.12s)
the past like 10 years or so we've not
[67:36] (4056.88s)
talked too much about software
[67:37] (4057.84s)
architecture. So maybe there is a little
[67:39] (4059.20s)
bit to learn from from other other
[67:40] (4060.80s)
people's ideas. So uh very to wrap up
[67:43] (4063.92s)
how about we just wrap up with some
[67:45] (4065.12s)
rapid questions. I I just asked a
[67:46] (4066.40s)
question. Okay, let's go for it. an
[67:47] (4067.76s)
issue. So first of all, what is your AI
[67:49] (4069.36s)
stack for coding? Cursor for hackathons,
[67:53] (4073.12s)
deep research to get a sense of what
[67:54] (4074.80s)
libraries already exist. Chatbt is my
[67:57] (4077.84s)
default search and also my tutor and
[67:59] (4079.60s)
some internal tools to quickly do rag
[68:01] (4081.76s)
over company documentations when I'm
[68:03] (4083.36s)
trying to find something. So what is a
[68:05] (4085.12s)
book that you would recommend and why?
[68:07] (4087.52s)
Book I'd recommend is the almanac of
[68:10] (4090.16s)
Neville Ravikant. I've read it a couple
[68:12] (4092.24s)
times. big fan of the way he talks about
[68:15] (4095.76s)
building your life both from a very
[68:17] (4097.84s)
pragmatic way about how you should do it
[68:19] (4099.44s)
from your career but also in terms of
[68:21] (4101.44s)
just how to be happy and what is a piece
[68:23] (4103.68s)
of advice that made a big difference in
[68:25] (4105.28s)
your professional career don't wait for
[68:27] (4107.04s)
someone to give you the opportunity to
[68:28] (4108.32s)
go work on something go work on it love
[68:30] (4110.32s)
it so John this was so nice to to have
[68:33] (4113.44s)
you on the show thank you so much for
[68:35] (4115.12s)
even having me in the first place thanks
[68:36] (4116.72s)
very much to John for this conversation
[68:38] (4118.64s)
to me talking with her was a great
[68:40] (4120.24s)
reminder of how in a new field like
[68:41] (4121.52s)
Genai years ago experience might be less
[68:43] (4123.52s)
relevant than teaching herself how to
[68:45] (4125.04s)
use these new technologies like Jambi
[68:47] (4127.04s)
has done. So, it's also a good reminder
[68:49] (4129.12s)
of how it's never too late to get
[68:50] (4130.56s)
started. Jambi thought that she was late
[68:52] (4132.96s)
in 2022 because she was 5 years behind
[68:56] (4136.24s)
every AI researcher who's been using
[68:58] (4138.00s)
transformers since it was released in
[69:00] (4140.20s)
2017. And yet, Jambi is now working at
[69:02] (4142.72s)
OpenAI, the company that arguably made
[69:04] (4144.96s)
the most in utilizing transformers and
[69:07] (4147.24s)
LLMs. For more in-depth deep dives on
[69:09] (4149.76s)
how OpenAI works coming from the OpenAI
[69:12] (4152.16s)
team and on practical guides on AI
[69:14] (4154.24s)
engineering, check out the pragmatic
[69:15] (4155.76s)
engineer deep dives which are linked in
[69:17] (4157.20s)
the show notes below. If you enjoy this
[69:19] (4159.04s)
podcast, please do subscribe on your
[69:20] (4160.88s)
favorite podcast platform and on
[69:22] (4162.48s)
YouTube. A special thank you if you
[69:24] (4164.56s)
leave a review which greatly helps the
[69:26] (4166.40s)
podcast. Thanks and see you in the next