1
00:00:00,000 --> 00:00:02,750
Bret: Welcome to DevOps and Docker
talk, and I'm your host, Bret.

2
00:00:03,630 --> 00:00:05,460
This episode is a special one.

3
00:00:05,520 --> 00:00:10,860
It's actually the first episode from a
totally new podcast I launched called

4
00:00:10,860 --> 00:00:15,930
Agentic DevOps, and that podcast is
gonna run in parallel with this one.

5
00:00:16,290 --> 00:00:19,650
So this one, the goal is
still for the last six years.

6
00:00:20,205 --> 00:00:24,795
Everything related to containers,
cloud native Kubernetes, and Docker,

7
00:00:24,795 --> 00:00:27,435
and the DevOps workloads around that.

8
00:00:27,975 --> 00:00:29,475
And I don't plan on changing any of that.

9
00:00:29,475 --> 00:00:31,455
We're gonna still have the same guests.

10
00:00:31,455 --> 00:00:36,135
A certain amount of those will be AI
related guests, but I was seeing a trend.

11
00:00:36,135 --> 00:00:38,379
I. That I'll talk about in the show.

12
00:00:38,739 --> 00:00:44,290
And I thought that Agentic DevOps was
going to be a big thing here in 2025.

13
00:00:44,290 --> 00:00:49,480
So a few months back we started working on
content episodes and theming and branding.

14
00:00:49,690 --> 00:00:55,720
A whole new podcast that I recommend
you check out at agenticdevops.Fm

15
00:00:56,000 --> 00:00:57,110
links in the show notes.

16
00:00:57,380 --> 00:01:00,680
And this is the first episode from
that podcast that I'm just presenting

17
00:01:00,680 --> 00:01:02,270
here so that you can check it out.

18
00:01:02,556 --> 00:01:07,446
Neral and I talk theory around what
we see coming and what might be

19
00:01:07,446 --> 00:01:12,126
a huge shift in how we use AI to
do our jobs as DevOps engineers.

20
00:01:12,516 --> 00:01:16,506
And that intention for that show is
to have more guests and to really

21
00:01:16,566 --> 00:01:20,466
dial in and focus on that very
niche topic, at least for this year.

22
00:01:20,466 --> 00:01:20,976
Who knows?

23
00:01:21,026 --> 00:01:23,306
it might be a bigger deal than this show,

24
00:01:23,556 --> 00:01:27,096
so if you enjoy this episode,
subscribe to that second podcast of

25
00:01:27,096 --> 00:01:29,556
mine, and now I'm gonna have two.

26
00:01:29,736 --> 00:01:30,756
So I hope you enjoy.

27
00:01:41,084 --> 00:01:45,824
Welcome to the first episode of
my new podcast, a Agentic DevOps.

28
00:01:45,964 --> 00:01:46,954
this episode.

29
00:01:47,674 --> 00:01:53,374
Is kicking off what I think is going
to be a big topic for my entire year,

30
00:01:53,554 --> 00:01:59,704
probably for the next few years around
wrangling AI into some usable format.

31
00:02:00,094 --> 00:02:06,384
For DevOps, you probably heard of AI
agents by now, or the MCP protocol.

32
00:02:06,504 --> 00:02:09,384
I guess I should just say MCP,
since P stands for protocol.

33
00:02:09,714 --> 00:02:14,454
And these two things together are
creating potentially something

34
00:02:14,454 --> 00:02:18,384
very useful for platform
engineering, DevOps, and that stuff.

35
00:02:18,684 --> 00:02:21,174
it has so much potential that.

36
00:02:21,564 --> 00:02:26,304
In the first quarter of 2025, I kind of
thought this was gonna be a big deal.

37
00:02:26,304 --> 00:02:29,754
This was gonna be, uh, if we can
figure out how to keep these things

38
00:02:29,754 --> 00:02:32,394
from hallucinating and going crazy
in our infrastructure, this could

39
00:02:32,394 --> 00:02:36,834
potentially be the AI shift for
infrastructure that I was waiting for.

40
00:02:37,254 --> 00:02:39,294
So started this podcast.

41
00:02:39,354 --> 00:02:43,749
We recorded our first episode at
KubeCon at the beginning of April,

42
00:02:43,749 --> 00:02:51,024
2025, and this is gonna be a series of
very specific episodes around getting.

43
00:02:52,089 --> 00:02:56,679
Ais to do useful automation and work
for DevOps, platform engineering,

44
00:02:56,859 --> 00:02:59,979
infrastructure management,
cloud, you know, all those things

45
00:03:00,309 --> 00:03:02,619
beyond just writing YAML, right?

46
00:03:02,889 --> 00:03:06,579
So the, the intro for this podcast,
there's a separate episode for intro.

47
00:03:06,579 --> 00:03:09,759
It kind of goes into my whole theory of
why I think this is gonna be a thing.

48
00:03:10,059 --> 00:03:13,599
And this episode we really try to break
down the basics and fundamentals for

49
00:03:13,599 --> 00:03:15,189
those of you that are catching up.

50
00:03:15,189 --> 00:03:16,524
Because it's a lot.

51
00:03:16,674 --> 00:03:17,994
There's a lot going on.

52
00:03:18,174 --> 00:03:22,924
It seems like We have announcements
every day this year around AI agents or

53
00:03:22,924 --> 00:03:24,954
Agentic, ai, however you wanna call it.

54
00:03:25,204 --> 00:03:29,224
I am calling it Agentic DevOps,
and hoping that name will stick.

55
00:03:29,674 --> 00:03:32,734
Now, this episode, since it's
from the beginning of April.

56
00:03:33,439 --> 00:03:37,309
And it is technically now just getting
released at the beginning of June.

57
00:03:37,819 --> 00:03:40,189
We're a little bit behind on
launching this new podcast.

58
00:03:40,519 --> 00:03:42,709
Um, I think everything
in it's still relevant.

59
00:03:42,709 --> 00:03:44,179
There's just been a lot more since.

60
00:03:44,389 --> 00:03:46,009
And I don't know the frequency yet.

61
00:03:46,009 --> 00:03:48,229
I don't know how often this
podcast is gonna happen.

62
00:03:48,469 --> 00:03:50,569
It could be potentially every other week.

63
00:03:50,749 --> 00:03:51,619
It could be weekly.

64
00:03:51,619 --> 00:03:54,649
I just don't know yet because we
are not gonna do the same thing

65
00:03:54,649 --> 00:03:56,719
here as on my usual podcast.

66
00:03:56,719 --> 00:03:59,509
If you're someone who knows that
one DevOps and Docker talk that I've

67
00:03:59,509 --> 00:04:02,869
been doing the last seven years, that
one is still gonna have AI in it.

68
00:04:02,869 --> 00:04:07,039
But this one is very specific and
there might be a few episodes that have

69
00:04:07,039 --> 00:04:11,629
syndication or whatever you wanna call
it, of the episodes on both podcasts.

70
00:04:12,184 --> 00:04:16,594
But most of the time we're gonna
keep the focus of just everything,

71
00:04:16,594 --> 00:04:20,434
DevOps, everything, containers on
the DevOps and Docker talk show.

72
00:04:20,494 --> 00:04:24,934
And this one is gonna be very
specific around implementing useful

73
00:04:25,564 --> 00:04:31,744
AI related things for Agentic DevOps,
or automating our DevOps with robots.

74
00:04:32,224 --> 00:04:35,734
So I hope you enjoyed this episode
with Nirmal from KubeCon London.

75
00:04:41,457 --> 00:04:42,187
Hey, I'm Bret.

76
00:04:42,257 --> 00:04:43,337
And we're at Kon.

77
00:04:44,282 --> 00:04:45,972
We are Hi, Nirmal.

78
00:04:46,292 --> 00:04:47,072
Nirmal: I'm Nirmal Metha.

79
00:04:47,072 --> 00:04:50,972
I'm a principal specialist solution
architect at AWS and these are

80
00:04:50,972 --> 00:04:55,472
my views and not of my employers,
but this episode is all about

81
00:04:55,562 --> 00:04:56,042
Bret: AI

82
00:04:56,042 --> 00:04:56,822
Nirmal: agents

83
00:04:57,122 --> 00:04:59,762
Bret: for DevOps and platform engineering.

84
00:04:59,762 --> 00:04:59,792
Ooh.

85
00:04:59,852 --> 00:05:03,332
So let's just start off real
quick with what is an AI agent?

86
00:05:03,362 --> 00:05:03,542
Okay.

87
00:05:03,542 --> 00:05:07,922
So we've heard of ai, we
know ai, gen, AI chat, GPT.

88
00:05:08,132 --> 00:05:09,002
We've talked about.

89
00:05:09,342 --> 00:05:12,672
running LLMs, running
inference on platforms.

90
00:05:12,702 --> 00:05:12,882
Yep.

91
00:05:12,942 --> 00:05:17,922
And that we are managing the workloads
that provide other people services.

92
00:05:17,982 --> 00:05:18,432
Absolutely.

93
00:05:18,702 --> 00:05:21,642
So how is AI agents different than that?

94
00:05:22,202 --> 00:05:24,742
Nirmal: This is a air in
terms of bleeding edge.

95
00:05:24,952 --> 00:05:25,192
Yeah.

96
00:05:25,672 --> 00:05:26,812
This is it, right?

97
00:05:26,812 --> 00:05:26,872
Yeah.

98
00:05:26,872 --> 00:05:28,132
Like we're a year ago.

99
00:05:28,132 --> 00:05:28,492
No one

100
00:05:28,492 --> 00:05:28,792
Bret: had this

101
00:05:28,792 --> 00:05:29,212
Nirmal: term

102
00:05:29,272 --> 00:05:29,902
Bret: six months ago.

103
00:05:29,902 --> 00:05:30,412
I don't think anybody's

104
00:05:30,412 --> 00:05:30,832
Nirmal: talking about it.

105
00:05:30,832 --> 00:05:31,417
I'm very few people.

106
00:05:31,487 --> 00:05:32,457
Yeah, very few people.

107
00:05:32,957 --> 00:05:36,977
and we've seen it in the news a
lot of vendors and big companies

108
00:05:37,027 --> 00:05:41,637
announcing Agentic ai, that's another
term's ai, so AI agents, Agentic

109
00:05:42,047 --> 00:05:49,937
It's giving your LLM, like your chat,
GPT or your Claude or local LM Lama.

110
00:05:49,997 --> 00:05:50,297
Yeah.

111
00:05:51,197 --> 00:05:55,157
Access to run commands.

112
00:05:55,257 --> 00:05:56,337
On your behalf.

113
00:05:56,697 --> 00:05:57,867
Or on its behalf.

114
00:05:58,437 --> 00:05:58,617
Bret: Yeah.

115
00:05:58,617 --> 00:06:01,047
And we call those tools like
that, if you hear that word.

116
00:06:01,107 --> 00:06:01,227
Tools.

117
00:06:01,227 --> 00:06:01,407
Yeah.

118
00:06:01,437 --> 00:06:04,137
That's like the generic
tool, like I guess a shell.

119
00:06:04,602 --> 00:06:05,532
Could be a tool.

120
00:06:05,562 --> 00:06:06,042
Correct.

121
00:06:06,092 --> 00:06:08,552
Reading a file could be a tool.

122
00:06:08,702 --> 00:06:12,572
Accessing a remote, API of
a web service is a tool.

123
00:06:12,602 --> 00:06:12,872
Yep.

124
00:06:12,902 --> 00:06:14,402
Searching could be a tool.

125
00:06:14,402 --> 00:06:18,322
And so these tools what what makes
that different than what we've

126
00:06:18,322 --> 00:06:20,962
been seeing in our code editors?

127
00:06:20,992 --> 00:06:21,322
Yeah.

128
00:06:21,472 --> 00:06:22,202
How is that different?

129
00:06:22,586 --> 00:06:25,636
Nirmal: I'm a platform engineer
and I want to build out an

130
00:06:25,636 --> 00:06:27,316
EKS cluster using Terraform.

131
00:06:27,316 --> 00:06:28,096
That's what we use.

132
00:06:28,396 --> 00:06:31,816
So I'll ask let's say Claude or chat GBT.

133
00:06:31,936 --> 00:06:32,116
Yeah.

134
00:06:32,176 --> 00:06:36,796
I'm a platform engineer and I want to
build a production ready EKS cluster.

135
00:06:36,826 --> 00:06:37,636
Please create.

136
00:06:37,936 --> 00:06:42,166
The assets I need, and it
will spit out some Terraform.

137
00:06:42,171 --> 00:06:42,436
Yaml, right?

138
00:06:42,436 --> 00:06:42,646
Yeah.

139
00:06:43,006 --> 00:06:44,231
Bret: And it's writing text.

140
00:06:44,231 --> 00:06:45,161
Nirmal: It's writing text.

141
00:06:45,161 --> 00:06:46,781
And I can, I'll double
you a little button.

142
00:06:46,781 --> 00:06:47,621
I copy that.

143
00:06:47,621 --> 00:06:50,351
Put it in, or there'll be, if you're
using Cursor, all these other tools,

144
00:06:50,351 --> 00:06:52,891
you can put it into some TF file.

145
00:06:52,891 --> 00:06:52,951
Yeah.

146
00:06:53,491 --> 00:06:57,001
I can then take that and I can ask
the LM what's the command that I

147
00:06:57,001 --> 00:07:00,121
need to run to apply this Terraform?

148
00:07:00,578 --> 00:07:04,876
To actually stand up the, what it's,
what's described in this terraform.

149
00:07:05,086 --> 00:07:07,786
It'll spit out, okay, you wanna
do Terraform plan and then

150
00:07:07,786 --> 00:07:09,016
Terraform apply and all that.

151
00:07:09,016 --> 00:07:12,731
Terraform in it or whatever, and
I'll just copy those commands and

152
00:07:12,761 --> 00:07:14,621
check 'em and write them myself.

153
00:07:15,671 --> 00:07:20,171
So the LLM is not executing
anything on my behalf.

154
00:07:20,171 --> 00:07:20,666
On, on your behalf.

155
00:07:21,031 --> 00:07:25,051
Agent would be defining a tool set.

156
00:07:25,111 --> 00:07:30,031
So I could give, I could define
a tool called Terraform or a tool

157
00:07:30,031 --> 00:07:35,881
called Shell I could describe what
that tool does in natural language.

158
00:07:36,181 --> 00:07:36,421
Bret: Okay.

159
00:07:36,721 --> 00:07:41,581
Nirmal: And then I can give
the LLM system a list of these

160
00:07:41,581 --> 00:07:43,111
tools and their descriptions.

161
00:07:43,411 --> 00:07:44,851
And tell it.

162
00:07:45,091 --> 00:07:45,751
Okay?

163
00:07:46,681 --> 00:07:47,731
Back to the same scenario.

164
00:07:47,731 --> 00:07:51,151
I'm a platform engineer and I want
to create an EKS production cluster

165
00:07:51,151 --> 00:07:56,199
using Terraform, and I want you
to create it right for me because

166
00:07:56,199 --> 00:07:58,149
it has the access to those tools.

167
00:07:58,149 --> 00:08:04,089
Now it internal reasons, okay,
I need to create some Terraform.

168
00:08:04,389 --> 00:08:07,089
I need to validate it in some
kind of way, and then I need.

169
00:08:07,639 --> 00:08:09,199
I need to execute this Terraform.

170
00:08:09,769 --> 00:08:11,629
Is there any tools that
I have in my toolbox

171
00:08:11,629 --> 00:08:13,759
Bret: In this case, sorry the
i is the, you're referring

172
00:08:13,759 --> 00:08:14,809
to yourself as the ai, right?

173
00:08:15,169 --> 00:08:15,319
Yeah.

174
00:08:15,319 --> 00:08:15,499
Sorry.

175
00:08:16,344 --> 00:08:17,809
It's no longer the
human doing this, right?

176
00:08:17,809 --> 00:08:17,893
No.

177
00:08:17,893 --> 00:08:19,303
We gave it instructions and we sit back

178
00:08:19,843 --> 00:08:22,277
Nirmal: from the perspective, from
the perspective of the, LLM the

179
00:08:22,277 --> 00:08:27,377
Gen AI tool itself, the LLM system
that's the I in this scenario.

180
00:08:27,377 --> 00:08:27,378
Yeah.

181
00:08:28,277 --> 00:08:30,137
I, the LLM is deciding.

182
00:08:31,007 --> 00:08:38,267
The Gen NI tool is looking at its list of
available tools and matching what it needs

183
00:08:38,267 --> 00:08:44,327
to it, figure it, it's reasoning about
what the end goal is and it looks and

184
00:08:44,327 --> 00:08:48,437
says, there's this tool called Terraform
that allows me to use infrastructure as

185
00:08:48,437 --> 00:08:51,167
code to deploy resources on the cloud.

186
00:08:51,677 --> 00:08:52,847
That sounds like what I need.

187
00:08:52,997 --> 00:08:53,507
Maybe.

188
00:08:53,647 --> 00:08:55,177
And it.

189
00:08:55,867 --> 00:08:58,567
Generates the terraform just like
it did the first time around.

190
00:08:59,137 --> 00:09:02,017
It knows what command to run.

191
00:09:02,017 --> 00:09:07,067
It generates the command and then
the magic here, a little box will

192
00:09:07,067 --> 00:09:10,907
show up and says, do you want me
to execute this on your behalf?

193
00:09:11,117 --> 00:09:14,897
You click the button, you click the
button, and then it executes that

194
00:09:14,927 --> 00:09:24,107
Terraform apply Uhhuh and it sounds very
simple, but it's a very different paradigm

195
00:09:24,107 --> 00:09:29,517
in terms of thinking about how we interact
with infrastructure or systems in general.

196
00:09:29,517 --> 00:09:31,227
Like broadly systems in general.

197
00:09:31,257 --> 00:09:35,247
Because we are no, like in this
way of looking at it or thinking

198
00:09:35,247 --> 00:09:42,057
about it, I, as the human, are no
longer executing those commands.

199
00:09:42,182 --> 00:09:42,272
I am.

200
00:09:42,302 --> 00:09:47,162
Trusting to a certain extent that
the LLM can figure out what it needs

201
00:09:47,162 --> 00:09:53,752
to do and giving it a guardrail
set of tools to use and execute.

202
00:09:54,052 --> 00:09:54,502
Bret: Yeah.

203
00:09:54,562 --> 00:09:57,212
And so we're giving the, we're
giving the Chaos monkey XI

204
00:09:57,212 --> 00:09:58,172
mean, it's automation, right?

205
00:09:58,172 --> 00:10:00,722
We could actually classify
this as just automation.

206
00:10:00,722 --> 00:10:01,622
It just happens to be.

207
00:10:02,357 --> 00:10:04,547
Figuring out what to
automate in real time.

208
00:10:04,787 --> 00:10:08,657
Rather than the traditional automation
where we have a very deterministic plan

209
00:10:08,657 --> 00:10:12,594
of, steps that are repeated over and
over again by a GitHub action runner

210
00:10:12,624 --> 00:10:14,544
or a CI CD platform or something.

211
00:10:14,754 --> 00:10:14,844
Yeah.

212
00:10:14,844 --> 00:10:20,544
Nirmal: And the agent part is the
piece of software that enables.

213
00:10:21,249 --> 00:10:22,359
The LLM to execute.

214
00:10:22,659 --> 00:10:22,929
Bret: Yeah.

215
00:10:22,979 --> 00:10:27,409
Nirmal: and pull, pulls this all
together and one, so back to what I was

216
00:10:27,409 --> 00:10:30,949
talking about with the infrastructure
and there was a part where I said,

217
00:10:30,979 --> 00:10:37,939
okay, how do we define what tools are
available for the agent system to use?

218
00:10:37,939 --> 00:10:42,761
and how do I want the
agent to call those tools?

219
00:10:43,946 --> 00:10:46,826
And reason about them, and
there's a protocol called

220
00:10:46,826 --> 00:10:48,896
MCP Model Context Protocol.

221
00:10:48,996 --> 00:10:54,306
Just outlining a standard way of
defining the tools, the system prompt

222
00:10:54,796 --> 00:10:56,176
for that tool and a description.

223
00:10:56,476 --> 00:10:58,836
Bret: And this is like an API where
you like define the spec of an API.

224
00:10:58,836 --> 00:11:04,081
Nirmal: It's a defined spec of an
API and the adoption of that API is

225
00:11:04,081 --> 00:11:05,101
Bret: just exploding right now,

226
00:11:05,221 --> 00:11:05,731
Nirmal: essentially.

227
00:11:05,821 --> 00:11:05,881
Bret: Yeah.

228
00:11:06,031 --> 00:11:08,971
So we're to, to under if you're not,
okay sorry, lemme back up a second.

229
00:11:09,211 --> 00:11:12,451
That's a very valid point because that's
the reason I wanted to record This's

230
00:11:12,451 --> 00:11:14,301
a I don't wanna be a hype machine.

231
00:11:14,631 --> 00:11:14,781
Correct.

232
00:11:14,781 --> 00:11:16,791
But I'm super excited right now.

233
00:11:16,951 --> 00:11:21,391
if you can see inside my, in
my enthusiastic brain, I've

234
00:11:21,391 --> 00:11:24,391
only been paying attention to
this for a little over a month.

235
00:11:24,706 --> 00:11:27,046
If you asked me two months ago
what an AI agent was, I'd say,

236
00:11:27,286 --> 00:11:29,266
I don't know a robot that's ai.

237
00:11:29,296 --> 00:11:29,791
I don't know.

238
00:11:30,436 --> 00:11:32,716
I now think I've got a
much better handle on this.

239
00:11:32,791 --> 00:11:36,556
I've been spending so much of my life
right now, deep diving into this, to

240
00:11:36,556 --> 00:11:39,586
the point that you and I are talking
about changing some of the focus

241
00:11:39,586 --> 00:11:41,026
this year on, on all these topics.

242
00:11:41,026 --> 00:11:41,386
Absolutely.

243
00:11:41,386 --> 00:11:44,056
Because I think this is gonna
dominate the conversation.

244
00:11:44,356 --> 00:11:46,846
This is, these are, there's gonna be
a lot of predictions in this and we're

245
00:11:46,846 --> 00:11:49,876
not gonna talk forever 'cause it's
gonna need to be multiple episodes to

246
00:11:49,876 --> 00:11:51,556
really break down what's going on here.

247
00:11:51,556 --> 00:11:52,906
But we now have the definitions.

248
00:11:53,156 --> 00:11:55,166
AI agents, what are tools?

249
00:11:55,586 --> 00:11:58,196
The protocol behind it is
essentially MCP right now.

250
00:11:58,196 --> 00:12:01,406
Although that's not necessarily gonna be
the only thing, it's just the thing right

251
00:12:01,406 --> 00:12:03,896
now that we're agreeing on by one company.

252
00:12:03,956 --> 00:12:04,466
Exactly.

253
00:12:04,766 --> 00:12:10,626
Nirmal: We have to caveat this with, this
is like this is early like Docker days.

254
00:12:10,656 --> 00:12:11,166
This is like

255
00:12:11,466 --> 00:12:14,016
Bret: Docker in day 60, right?

256
00:12:14,016 --> 00:12:14,106
Yes.

257
00:12:14,106 --> 00:12:17,726
Like we were like right after
Python in 2013 when we gave that

258
00:12:17,726 --> 00:12:19,136
de, when he gave that demo, Solomon.

259
00:12:19,616 --> 00:12:23,966
Like we all saw it and didn't
understand it fully, but it

260
00:12:23,966 --> 00:12:25,316
felt like something right.

261
00:12:25,316 --> 00:12:28,886
And like you and I both, that's why
we were early docker captains, is

262
00:12:28,886 --> 00:12:31,586
we saw that as a platform shift.

263
00:12:31,804 --> 00:12:35,294
we've seen these waves before over,
over our careers of many decades

264
00:12:35,484 --> 00:12:40,064
that we earned with this gray
beard status with effort and toil.

265
00:12:40,364 --> 00:12:43,814
And I feel like this is maybe the moment.

266
00:12:44,624 --> 00:12:48,794
That was the moment of 2013 and that,
and yeah, I'm not alone in that feeling.

267
00:12:48,854 --> 00:12:49,134
yes.

268
00:12:49,134 --> 00:12:52,674
Nirmal: And there's just to be clear,
there's massive differences between

269
00:12:52,794 --> 00:12:57,024
like paradigm shifts in terms of like
virtualization, cloud containers.

270
00:12:57,104 --> 00:13:02,004
And the tooling of software
development and systems development

271
00:13:02,004 --> 00:13:05,964
and right systems operations,
it's still in that same vein, but.

272
00:13:06,324 --> 00:13:07,154
Yeah, we're not replacing,

273
00:13:07,154 --> 00:13:09,944
Bret: this is not replacing infrastructure
or containers or anything like that.

274
00:13:09,994 --> 00:13:11,914
This is just gonna change the way we work.

275
00:13:12,214 --> 00:13:12,634
Nirmal: Correct.

276
00:13:12,634 --> 00:13:15,724
And also it's broader than
just like IT infrastructure.

277
00:13:15,854 --> 00:13:20,624
Like this has implications with
software design or application,

278
00:13:20,624 --> 00:13:22,154
like what an application does.

279
00:13:22,254 --> 00:13:25,344
And I want to think of
this as a teaser trailer.

280
00:13:25,659 --> 00:13:27,644
To subsequent new series, episode.

281
00:13:27,644 --> 00:13:27,766
A new series.

282
00:13:27,771 --> 00:13:28,449
Yeah, absolutely.

283
00:13:28,479 --> 00:13:28,779
We're gonna have to

284
00:13:28,779 --> 00:13:29,589
Bret: come up with a name.

285
00:13:29,589 --> 00:13:32,739
I'm toying around with the idea of
Agentic DevOps, and just classifying

286
00:13:32,739 --> 00:13:36,709
that as the absolutely as the theme
of certain levels of podcast episodes.

287
00:13:36,709 --> 00:13:37,760
You've heard it here first.

288
00:13:37,760 --> 00:13:37,904
Heard it here first.

289
00:13:37,939 --> 00:13:38,139
This

290
00:13:38,169 --> 00:13:39,439
Nirmal: is Agentic DevOps.

291
00:13:39,509 --> 00:13:41,999
Another term we're seeing is AI four ops.

292
00:13:42,299 --> 00:13:43,559
Again, this is early days.

293
00:13:43,589 --> 00:13:44,549
None of this is like

294
00:13:44,609 --> 00:13:44,789
Bret: Yeah.

295
00:13:44,789 --> 00:13:45,569
Set in stone at all.

296
00:13:45,569 --> 00:13:48,839
Yeah, and if you're at KU Con today
with us, if you were here at this

297
00:13:48,839 --> 00:13:52,019
conference all week, AI was a constant
topic, but it wasn't about this.

298
00:13:52,394 --> 00:13:57,164
It actually, there was only one talk in
an entire week that even touched on the

299
00:13:57,164 --> 00:14:03,794
idea of using AI to do the job of an
DevOps or operator or platform engineer.

300
00:14:03,854 --> 00:14:06,464
Like people are, what we're talking
about at KU Con for the last three

301
00:14:06,464 --> 00:14:10,664
years has been how to run the
inference and build the LLM models.

302
00:14:11,054 --> 00:14:15,164
And so we are just still using
human effort to do that work.

303
00:14:15,359 --> 00:14:19,514
But this, I feel like I'm gonna draw the
line in the sand and say, this is the.

304
00:14:19,974 --> 00:14:24,804
month or the definitely
the year, that kicks off.

305
00:14:25,274 --> 00:14:30,214
What will be a multi-year effort of
figuring out how we use automated

306
00:14:30,214 --> 00:14:33,874
LLMs Essentially with access to all
the tools we want to give it with

307
00:14:33,904 --> 00:14:36,284
the proper permissions and only
the permissions we want to give

308
00:14:36,284 --> 00:14:39,074
it right to do our work for us.

309
00:14:39,479 --> 00:14:42,809
In a less chaos mon monkey way, right?

310
00:14:42,809 --> 00:14:44,009
Like less chaotic way.

311
00:14:44,014 --> 00:14:44,204
Potentially.

312
00:14:44,279 --> 00:14:44,639
Potentially.

313
00:14:44,729 --> 00:14:46,709
It could, this thing can
easily go off the rails.

314
00:14:46,799 --> 00:14:47,189
Absolutely.

315
00:14:47,189 --> 00:14:51,959
I will probably reference in the show
notes Solomon Hike's recent talks about

316
00:14:51,969 --> 00:14:56,329
how they're now using Dagger, which
is primarily A-C-I-C-D pipeline tool.

317
00:14:56,379 --> 00:15:01,179
So he's talking, and a lot of my
language is actually from him iterating

318
00:15:01,179 --> 00:15:04,899
on his idea of what this might look
like when we're throwing a bunch

319
00:15:04,899 --> 00:15:10,599
of crazy hallucinating AI into what
we consider a deterministic world.

320
00:15:10,969 --> 00:15:11,219
Correct.

321
00:15:11,219 --> 00:15:17,399
Nirmal: I think with containers and cloud
and on the infrastructure APIs we have.

322
00:15:17,894 --> 00:15:22,694
We were chipping away and really
aiming at deterministic behavior

323
00:15:22,694 --> 00:15:24,566
with respect to infrastructure.

324
00:15:26,434 --> 00:15:28,646
Ironically, maybe not
ironically, I don't know.

325
00:15:29,006 --> 00:15:33,776
Now we're introducing a paradigm
shift that reintroduces a lot

326
00:15:33,776 --> 00:15:35,971
of non-determinism right into.

327
00:15:36,656 --> 00:15:40,376
A place that we have been fighting
to non-determinism for a long time.

328
00:15:41,066 --> 00:15:43,376
Bret: We have been working
to get rid of all that.

329
00:15:43,376 --> 00:15:46,226
And now we're, that's why I keep
saying Chaos monkey, because we're

330
00:15:46,226 --> 00:15:47,756
throwing a wrench into the system.

331
00:15:47,766 --> 00:15:52,056
That is in some ways feels like we're
going back to a world of, I don't

332
00:15:52,056 --> 00:15:53,706
know, what's the status of the system?

333
00:15:53,706 --> 00:15:54,036
I don't know.

334
00:15:54,706 --> 00:15:57,896
and this will probably be another
episode, I feel like this Agentic

335
00:15:57,896 --> 00:16:00,896
approach where we're actually
can have the potential to pit.

336
00:16:01,271 --> 00:16:03,341
The LLMs against each other, right?

337
00:16:03,341 --> 00:16:05,231
And have different
personas of these agents.

338
00:16:05,381 --> 00:16:07,901
One is the validator, one is the tester.

339
00:16:08,051 --> 00:16:10,121
One is one is the builder.

340
00:16:10,181 --> 00:16:11,771
And they can fight amongst each other.

341
00:16:11,861 --> 00:16:12,701
And it all works out.

342
00:16:12,701 --> 00:16:15,011
It actually ha happens to
actually work out better.

343
00:16:15,251 --> 00:16:19,091
And so if you're like me and
for the last three years of

344
00:16:19,151 --> 00:16:21,041
understanding, ever since GPT.

345
00:16:21,551 --> 00:16:22,961
3.5 or whatever came out.

346
00:16:22,961 --> 00:16:27,671
We all saw chat GPT as a product, and
then we started with GoodHub copilot and

347
00:16:27,721 --> 00:16:32,571
we started down this road As a DevOps
person, I haven't had a lot to talk about

348
00:16:32,811 --> 00:16:36,531
because I'm not interested in which model
is the fastest or the most accurate.

349
00:16:36,531 --> 00:16:37,191
'cause you know what?

350
00:16:37,321 --> 00:16:41,501
they all hallucinate and
still even today, years later.

351
00:16:42,491 --> 00:16:45,791
Code agents and we and you can see
this on YouTube, you watch basically

352
00:16:45,791 --> 00:16:49,421
thousands of videos on YouTube of
people trying to use these models to

353
00:16:49,421 --> 00:16:51,911
write perfect code and they just don't.

354
00:16:52,401 --> 00:16:56,151
And so we in ops, but we look at
that, I think, and the people I

355
00:16:56,151 --> 00:17:00,201
talk to even for years now are like,
we're never gonna use that for ops.

356
00:17:00,241 --> 00:17:03,061
But now my opinion has changed.

357
00:17:03,121 --> 00:17:03,271
Yeah.

358
00:17:03,621 --> 00:17:03,981
Nirmal: yeah.

359
00:17:03,981 --> 00:17:07,981
And I. If you're listening to this
and your gut reaction is, wait we

360
00:17:07,981 --> 00:17:10,291
have like APIs that are deterministic.

361
00:17:10,291 --> 00:17:10,891
Like you just

362
00:17:11,221 --> 00:17:11,461
Bret: Yeah.

363
00:17:11,761 --> 00:17:12,871
Nirmal: We can just call an API.

364
00:17:12,871 --> 00:17:17,431
We can have an automation tool call an
API to stand up infrastructure and like,

365
00:17:17,431 --> 00:17:23,041
why do we need to recreate like another
layer that makes it non-deterministic.

366
00:17:23,041 --> 00:17:27,751
And looks like an API but isn't an API
and you don't really know what it might

367
00:17:27,751 --> 00:17:30,481
do or which direction it might go.

368
00:17:30,841 --> 00:17:31,021
Yeah.

369
00:17:31,051 --> 00:17:32,261
And you're feeling I don't know.

370
00:17:32,261 --> 00:17:35,291
That doesn't seem like it would
solve any problems for me.

371
00:17:35,291 --> 00:17:37,301
And it seems like it might
introduce a lot of problems.

372
00:17:37,501 --> 00:17:39,811
You're in the right place because
that's exactly what we're gonna explore.

373
00:17:40,111 --> 00:17:40,531
Bret: Yeah.

374
00:17:40,881 --> 00:17:44,181
Nirmal: one thing for sure
though is it's here, right?

375
00:17:44,616 --> 00:17:51,206
I and so I feel like as good engineers,
as good system admins and operators

376
00:17:51,686 --> 00:17:53,301
Bret: are we enjoy, we love our crafts.

377
00:17:53,301 --> 00:17:54,231
We, we look at this as an.

378
00:17:54,546 --> 00:17:57,446
Art form of brain power and Right.

379
00:17:57,546 --> 00:18:00,766
Reaching for perfectionism in our
YAML and in our infrastructure

380
00:18:00,766 --> 00:18:02,716
optimization and our security.

381
00:18:02,936 --> 00:18:07,946
Nirmal: And we have a healthy
sense of skepticism on new tools,

382
00:18:07,946 --> 00:18:09,746
new processes, new mechanisms.

383
00:18:09,746 --> 00:18:09,956
Yeah.

384
00:18:10,016 --> 00:18:13,946
When you, when availability of your
services is paramount and reliability,

385
00:18:14,306 --> 00:18:16,796
you want to introduce new things in a.

386
00:18:17,186 --> 00:18:18,476
In a prudent manner.

387
00:18:18,576 --> 00:18:22,446
And so we're gonna take that
approach, but we're not going

388
00:18:22,446 --> 00:18:24,936
to dismiss that this exists.

389
00:18:24,991 --> 00:18:30,431
Clearly there's a lot of interest,
energy integration happening,

390
00:18:30,761 --> 00:18:35,781
experimentation happening and some
people are already starting to see value.

391
00:18:36,021 --> 00:18:36,201
Yeah.

392
00:18:36,251 --> 00:18:39,175
and we're gonna explore
with you where that, goes.

393
00:18:39,205 --> 00:18:39,415
Bret (2): Yeah.

394
00:18:39,415 --> 00:18:46,585
This, just to be clear, this is
KubeCon April, 2025 and almost

395
00:18:46,585 --> 00:18:48,565
no one is talking about this yet.

396
00:18:48,845 --> 00:18:52,835
It feels like it's right under the
surface of a lot of conversations and

397
00:18:52,835 --> 00:18:56,075
a lot of people maybe are thinking
about it, but I'm not even sure that

398
00:18:56,075 --> 00:18:58,880
we're honest with ourselves around.

399
00:18:59,645 --> 00:19:02,135
That this is coming,
whether we like it or not.

400
00:19:02,345 --> 00:19:09,455
And only because, yeah, not only, but
one of the large reasons is business.

401
00:19:10,055 --> 00:19:10,385
Okay.

402
00:19:10,745 --> 00:19:11,255
Lemme back up.

403
00:19:11,255 --> 00:19:15,275
You know how in a lot of organizations,
Kubernetes became a mandate, right?

404
00:19:15,275 --> 00:19:18,495
So there's lots of stories that came
out over the course of Kubernetes

405
00:19:18,495 --> 00:19:22,425
lifetime of teams being told that
they need to implement Kubernetes.

406
00:19:22,475 --> 00:19:27,545
It didn't come from a systems engineering
approach of solving a known problem.

407
00:19:27,545 --> 00:19:28,595
It came down.

408
00:19:28,880 --> 00:19:33,390
Because an executive decided that
they read a CIO magazine article

409
00:19:33,390 --> 00:19:35,580
that said Kubernetes was a cool
new thing and they did it right.

410
00:19:35,730 --> 00:19:37,020
I hear this all the time.

411
00:19:37,020 --> 00:19:41,280
I confirm this multiple times this
week with other people, and I now feel

412
00:19:41,280 --> 00:19:43,770
like we're not talking about it yet.

413
00:19:44,190 --> 00:19:51,300
But I did hear multiple analysts say their
organizations that they're working with

414
00:19:51,330 --> 00:19:57,480
expect that we are going to reduce the
number of personnel in infrastructure.

415
00:19:57,780 --> 00:19:58,830
Because of ai.

416
00:19:58,955 --> 00:20:02,315
the only way that's possible
is if we use agents to our

417
00:20:02,315 --> 00:20:04,295
advantage, because we can't, yeah.

418
00:20:04,295 --> 00:20:06,425
I still don't believe
we're replacing ourselves.

419
00:20:06,755 --> 00:20:09,735
I don't think the agents will
ever in, in the near term.

420
00:20:09,735 --> 00:20:12,955
And as far as we can see out, let's
say five years they will, they

421
00:20:12,955 --> 00:20:16,255
won't be running all infrastructure
in the world by themselves.

422
00:20:16,375 --> 00:20:17,815
They can't turn on servers.

423
00:20:17,995 --> 00:20:22,095
They maybe you can actually pixie boot
and do a power on a POE or whatever, but.

424
00:20:22,845 --> 00:20:27,405
Like we still need someone to give them
orders and rules and guidelines to go

425
00:20:27,405 --> 00:20:31,695
do the work, but to me, I'm starting
to wonder if very quickly, especially

426
00:20:31,695 --> 00:20:35,475
for those bleeding organizations that
are looking to squeeze out every cost

427
00:20:35,475 --> 00:20:40,425
optimization they can of their staff,
that they're going to be mandated to

428
00:20:40,605 --> 00:20:46,785
not just take AI as a code gen for yaml,
but to start using these agents to.

429
00:20:47,280 --> 00:20:51,375
Increase the velocity of their
work . And my, one of my stories is

430
00:20:51,375 --> 00:20:55,185
over the last 30 years I do this in
talks is every major shift has been

431
00:20:55,185 --> 00:20:57,675
about speed, cost reduction in speed.

432
00:20:57,915 --> 00:20:59,835
Sometimes we get 'em
both at the same time.

433
00:20:59,895 --> 00:21:01,275
Sometimes they're one or the other.

434
00:21:01,275 --> 00:21:03,435
We get a cost reduction, but we
don't go any faster, which is

435
00:21:03,435 --> 00:21:06,855
fine, or we're going faster, but
it's not necessarily cheaper yet.

436
00:21:06,855 --> 00:21:07,215
Nirmal: Right.

437
00:21:07,515 --> 00:21:07,755
Bret: And.

438
00:21:09,060 --> 00:21:13,020
I feel like this is maybe the next
one where We're gonna be feeling the

439
00:21:13,020 --> 00:21:17,010
pressure because all the devs are
gonna be writing code with ai, which

440
00:21:17,010 --> 00:21:21,150
in theory is going to improve their
performance, which means they're writing

441
00:21:21,150 --> 00:21:24,390
more code, shipping more, or need, or
wanting to ship more code, potentially.

442
00:21:24,390 --> 00:21:27,300
And if we're not using AI ourselves.

443
00:21:27,600 --> 00:21:32,250
To automate more of these platform
designs, platform build outs,

444
00:21:32,310 --> 00:21:35,430
troubleshooting when we're in production
and things are problematic and we

445
00:21:35,430 --> 00:21:38,050
don't wanna spend three hours trying
to find the source of the problem.

446
00:21:38,290 --> 00:21:43,510
If we're not starting to use agents to,
to automate a lot of that and reduce the

447
00:21:43,510 --> 00:21:48,550
time to market, so to speak, for a certain
feature or platform feature then I don't

448
00:21:48,550 --> 00:21:52,860
think these teams are gonna hire more
of us to help enable the devs to deploy.

449
00:21:53,190 --> 00:21:57,180
What it could end up happening is we
end up more with more shadow ops, where

450
00:21:57,180 --> 00:22:01,020
the developers are so fed up with us
not speeding up to the, if they're

451
00:22:01,020 --> 00:22:03,210
gonna go 10 x we have to go 10 x. Yeah.

452
00:22:03,260 --> 00:22:05,990
If they're gonna go three x or whatever
the number ends up being in the reports.

453
00:22:05,990 --> 00:22:09,255
And Gartner puts out like the AI
makes it efficient, more efficient

454
00:22:09,255 --> 00:22:11,235
for developers to, to code with ai.

455
00:22:11,235 --> 00:22:13,725
And the models get better and
the way they use it is better.

456
00:22:14,115 --> 00:22:17,575
And so they're shipping code faster and
they can do the same speed with three

457
00:22:17,575 --> 00:22:19,383
times less developers, or they can just.

458
00:22:19,825 --> 00:22:22,993
Produce three times more work, which I
think is more likely, because if it's

459
00:22:22,993 --> 00:22:25,903
the common denominator and everyone
has it, then that means every company

460
00:22:25,903 --> 00:22:29,323
can execute faster and they're gonna,
they're gonna want to do that because

461
00:22:29,323 --> 00:22:30,433
their competitors are doing that.

462
00:22:30,433 --> 00:22:33,193
So that's a's, that's a very
loaded and long prediction.

463
00:22:34,063 --> 00:22:35,173
Nirmal: That's a hypothesis.

464
00:22:35,663 --> 00:22:36,753
It's, I think there's
a lot of predict here.

465
00:22:36,753 --> 00:22:40,473
It's gonna take some time for us to
even chip away at that hypothesis,

466
00:22:40,473 --> 00:22:42,033
but it's a good starting point.

467
00:22:42,133 --> 00:22:47,593
If we're, but assuming that is like
the hypothesis that organizations

468
00:22:47,593 --> 00:22:51,103
are looking at to adopt these
tools that's a great starting point

469
00:22:51,103 --> 00:22:53,773
for us to help you figure out.

470
00:22:54,283 --> 00:22:57,133
what they are, why they are, what they do.

471
00:22:57,163 --> 00:22:57,313
Yeah.

472
00:22:57,313 --> 00:22:58,063
And how to use them.

473
00:22:58,363 --> 00:23:01,093
Bret: This is this, by the way, a
lot a little bit of that opinion

474
00:23:01,093 --> 00:23:03,553
of mine, and there's more to come
'cause I've got a lot more written

475
00:23:03,553 --> 00:23:04,423
down than we're never gonna get to.

476
00:23:04,843 --> 00:23:09,443
But a significant portion of that is
actually coming from what I've learned

477
00:23:09,443 --> 00:23:14,543
this week from analyst whose job it
is to figure this stuff out for their

478
00:23:14,543 --> 00:23:16,013
organization and their customers.

479
00:23:16,013 --> 00:23:16,373
Interesting.

480
00:23:16,553 --> 00:23:20,693
And so I, I am a little weighted by their.

481
00:23:21,698 --> 00:23:25,508
Almost unrealistic expectations
of how fast we can do this.

482
00:23:25,508 --> 00:23:26,678
'cause we are still humans.

483
00:23:26,838 --> 00:23:30,078
An organization can't adopt AI until
the humans learn how to adopt AI and

484
00:23:30,078 --> 00:23:31,728
the humans have to go at human speed.

485
00:23:31,808 --> 00:23:34,928
So we can't just flip a switch
and suddenly AI is here and

486
00:23:34,928 --> 00:23:35,888
running everything for us.

487
00:23:35,888 --> 00:23:38,978
At least not until we
have Iron Man's Jarvis.

488
00:23:39,008 --> 00:23:39,488
Or whatever.

489
00:23:39,488 --> 00:23:42,908
Like until we have that, we still have
to learn these tools and still have

490
00:23:42,908 --> 00:23:44,528
to adapt our platforms to use them.

491
00:23:44,528 --> 00:23:44,529
Yes.

492
00:23:44,534 --> 00:23:46,088
And adapt our learning to use them.

493
00:23:46,448 --> 00:23:47,738
And that's gonna take some time

494
00:23:47,798 --> 00:23:48,308
Nirmal: and.

495
00:23:48,663 --> 00:23:51,633
I'd like to, like the parting
thought for this is Okay.

496
00:23:51,633 --> 00:23:56,193
And here, like you said, there's an under
the surface kind of thing happening.

497
00:23:56,223 --> 00:23:56,583
Yeah.

498
00:23:56,943 --> 00:23:57,783
So whispers,

499
00:23:57,783 --> 00:23:58,983
Bret: it's almost like murmurs and under

500
00:23:58,983 --> 00:23:59,583
Nirmal: the surface.

501
00:23:59,588 --> 00:23:59,788
Yeah.

502
00:24:00,088 --> 00:24:02,793
AI agent, AI agents, mag

503
00:24:03,093 --> 00:24:03,423
Bret: DevOps.

504
00:24:03,453 --> 00:24:03,843
Ooh.

505
00:24:04,383 --> 00:24:05,583
This is our ASMR podcast.

506
00:24:05,583 --> 00:24:06,333
Moment of the podcast.

507
00:24:08,288 --> 00:24:09,518
Nirmal: Like MCP protocol.

508
00:24:09,818 --> 00:24:10,148
Bret: Yeah.

509
00:24:10,488 --> 00:24:14,448
Nirmal: you mentioned HA proxy on
the previous podcast, about load

510
00:24:14,448 --> 00:24:18,408
balancing and figuring out the
street, like token utilization of

511
00:24:18,408 --> 00:24:20,958
GPUs and tokens and all that stuff.

512
00:24:21,208 --> 00:24:25,318
and we had a conversation at the solo
booth and they were talking about having.

513
00:24:25,778 --> 00:24:30,008
A proxy for an MCP gateway, one of
the things that we're seeing the early

514
00:24:30,008 --> 00:24:32,688
signs of is these new workloads, right?

515
00:24:32,688 --> 00:24:39,268
This agentic kind of thinking Around
even just executing the agentic platform,

516
00:24:39,268 --> 00:24:43,678
if you will, And everything from
looking at the tokens and optimizing

517
00:24:43,678 --> 00:24:50,268
load balancing to inference endpoints
or MCP is, doesn't behave the same

518
00:24:50,268 --> 00:24:52,398
way as like just an http connection.

519
00:24:52,498 --> 00:24:53,158
Necessarily.

520
00:24:53,398 --> 00:24:53,968
And solo.

521
00:24:53,968 --> 00:24:56,368
We were talking to them and
they have an MCP gateway.

522
00:24:56,648 --> 00:24:59,198
We're seeing a little bit more
of a trend on AI gateways.

523
00:24:59,198 --> 00:25:04,598
Is DO the project has an AI gateway and
so this is not just another workload

524
00:25:04,778 --> 00:25:06,368
and looks like just a web server.

525
00:25:06,417 --> 00:25:09,117
And the networking and
everything is gonna be different.

526
00:25:09,197 --> 00:25:12,497
Not dramatically different,
but We'll, but drift different

527
00:25:12,497 --> 00:25:14,177
enough that we need to be aware.

528
00:25:14,777 --> 00:25:18,377
'cause even if you're not using
any of these tools, someone in your

529
00:25:18,377 --> 00:25:21,077
organization is probably gonna say,
oh, we need to integrate this stuff

530
00:25:21,107 --> 00:25:23,567
into our software, to our right.

531
00:25:23,567 --> 00:25:24,947
Whatever we're delivering.

532
00:25:25,427 --> 00:25:27,407
And we'll need to know
it even at that layer.

533
00:25:27,467 --> 00:25:30,887
So we're gonna also cover that
component as it relates to.

534
00:25:31,427 --> 00:25:32,897
The Kubernetes ecosystem, right?

535
00:25:32,912 --> 00:25:33,542
And cloud native.

536
00:25:33,842 --> 00:25:34,202
Bret: Yeah.

537
00:25:34,322 --> 00:25:38,172
I think this, if we had to do like an
elevator pitch for this podcast, it would

538
00:25:38,172 --> 00:25:46,252
be we now have a industry idea around
these terms agent, and then it uses an API

539
00:25:46,252 --> 00:25:50,568
called MCP to allow us to give more work.

540
00:25:50,898 --> 00:25:55,218
To these crazy robot texting things
that we have to talk to in human

541
00:25:55,218 --> 00:25:56,898
language and not with code, right?

542
00:25:56,898 --> 00:25:58,878
It's running code, but we're
not talking to it with code.

543
00:25:59,148 --> 00:26:04,238
And that it can now understand all the
tools we need to use and we can just give

544
00:26:04,238 --> 00:26:05,558
it a list of everything I wanted to use.

545
00:26:05,700 --> 00:26:08,820
here's my Kubernetes API, here's
all my other things that I, you have

546
00:26:08,820 --> 00:26:11,070
access to, and here's my problem.

547
00:26:11,400 --> 00:26:12,420
Go solve it.

548
00:26:12,570 --> 00:26:15,330
And that paradigm.

549
00:26:15,705 --> 00:26:19,125
Three months ago, two months ago
for me, I didn't know existed.

550
00:26:19,695 --> 00:26:22,695
And that's why I've been sitting
on the sidelines with ai.

551
00:26:22,695 --> 00:26:26,775
Like it's cool for writing programs
that mostly work in a demo.

552
00:26:27,025 --> 00:26:30,235
It's cool for adding a feature to
something I already have, but it's

553
00:26:30,235 --> 00:26:34,135
not doing my job as a platform
engineer or DevOps engineer.

554
00:26:34,185 --> 00:26:36,075
It's just helping me write text faster

555
00:26:36,169 --> 00:26:37,459
Then I can type into my keyboard.

556
00:26:37,509 --> 00:26:39,129
And that was not that interesting.

557
00:26:39,129 --> 00:26:41,499
That's why you didn't see a lot of
me talking about that on this show,

558
00:26:41,679 --> 00:26:42,609
was it just wasn't that interesting.

559
00:26:42,759 --> 00:26:48,369
This is an interesting topic for ops and
for absolutely engineers on the platform.

560
00:26:48,669 --> 00:26:48,849
Nirmal: Yep.

561
00:26:49,164 --> 00:26:49,464
Bret: So

562
00:26:49,944 --> 00:26:50,544
Nirmal: stay tuned.

563
00:26:50,634 --> 00:26:51,174
Yeah.

564
00:26:51,174 --> 00:26:54,504
And I, I love crazy texting robots.

565
00:26:54,534 --> 00:26:54,894
Crazy

566
00:26:54,894 --> 00:26:55,824
Bret: texting robots.

567
00:26:55,954 --> 00:26:57,544
Maybe that's the title.

568
00:26:57,544 --> 00:26:58,024
TBD.

569
00:27:00,244 --> 00:27:00,784
Alright.

570
00:27:00,844 --> 00:27:01,354
Alright.

571
00:27:01,354 --> 00:27:02,104
See you soon, man.

572
00:27:02,704 --> 00:27:02,914
See

573
00:27:02,914 --> 00:27:03,094
Nirmal: you.

574
00:27:03,094 --> 00:27:03,096
See you.

575
00:27:03,754 --> 00:27:04,324
Bye.

576
00:27:04,324 --> 00:27:04,384
Bye.