Practical AI

In this start-of-year FC episode, Chris and Daniel break down what really mattered in AI in 2025, and what to expect in 2026. They explore the rise of AI agents, the practical reality of multimodal AI, and how reasoning models are reshaping workflows. The conversation dives into infrastructure and energy constraints, the continued value of predictive models, and why orchestration (not just better models) is becoming the defining skill for AI teams. The episode wraps with grounded 2026 predictions on where AI systems, tooling, and builders are headed next.

Featuring:
Sponsor:
  • Framer - The enterprise-grade website builder that lets your team ship faster. Get 30% off at framer.com/practicalai
Upcoming Events: 

Creators and Guests

Host
Chris Benson
Cohost @ Practical AI Podcast • AI / Autonomy Research Engineer @ Lockheed Martin
Host
Daniel Whitenack

What is Practical AI?

Making artificial intelligence practical, productive & accessible to everyone. Practical AI is a show in which technology professionals, business people, students, enthusiasts, and expert guests engage in lively discussions about Artificial Intelligence and related topics (Machine Learning, Deep Learning, Neural Networks, GANs, MLOps, AIOps, LLMs & more).

The focus is on productive implementations and real-world scenarios that are accessible to everyone. If you want to keep up with the latest advances in AI, while keeping one foot in the real world, then this is the show for you!

Jerod:

Welcome to the Practical AI Podcast, where we break down the real world applications of artificial intelligence and how it's shaping the way we live, work, and create. Our goal is to help make AI technology practical, productive, and accessible to everyone. Whether you're a developer, business leader, or just curious about the tech behind the buzz, you're in the right place. Be sure to connect with us on LinkedIn, X, or Blue Sky to stay up to date with episode drops, behind the scenes content, and AI insights. You can learn more at practicalai.fm.

Jerod:

Now, onto the show.

Daniel:

Welcome to a new year of practical AI and an episode with just Chris and I where, we try to keep you fully connected with everything that's happening in the AI world, which is a lot these days, both last year and this year. But, I'm Daniel Witenack. I am CEO at Prediction Guard, and I'm joined as always by my cohost, Chris Benson, who is a principal AI research engineer at Lockheed Martin. Happy New Year, Chris.

Chris:

Hey. Happy New Year, Daniel. It's, 2026, probably the fastest moving AI year ever coming up here.

Daniel:

First, well, every new year, I guess, has been the fastest AI movie. Well, I guess since we started the podcast, you know, whatever, eight years ago.

Chris:

It was a safe thing for me to say. There you go.

Daniel:

Yeah, yeah, yeah. Safe thing for you to say. I mean, granted these last few years have been a little bit frantic, in relation to the years prior to that with the podcast, which felt, you know, in retrospect, seem a little bit chill. Yeah. But, it, it definitely seems like 2025 was a big year, 2026 will be a big year.

Daniel:

And so as we're coming into the new year for our listeners, usually we try to do some type of, we don't have a strict format here because we're pretty casual, but some type of discussion of things that happened in 2026, you know, themes looking forward or things that happened in 2025. I'm already a year ahead, I guess. Things that happened in 2025 and things that may or may not happen in 2026. Usually, our predictions are wrong as are all predictions.

Chris:

I'm okay with that.

Daniel:

All models are wrong, but hopefully this hopefully this podcast will be useful. Yeah. So interesting, interesting times, Chris, interesting dynamics in our world in all sorts of ways. But if we hone in on AI, I think at the beginning of last year, if I'm remembering correctly, there were a couple of things we talked about. One of those things, certainly at least a theme that we've talked about a lot this year, which if we were to categorize the year 2025, I don't know if you would agree, Chris, but it does seem like the year that we transitioned to talk about AI agents.

Daniel:

It was sort of like for a while we talked about models. Yeah. And then we kind of talked about assistance. And then we really kind of transitioned to talking about agents. Agents are autonomous AI.

Daniel:

That was a key theme of 2025. I guess one first question, Chris, did we actually What do we actually do with AI agents in 2025? Overall, was a positive and or successful year of trying agents?

Chris:

Well, I think I think there was, like, untold levels of hype around agents as there always is every time we hit a new thing. And I think a lot of I think a lot of organizations did try to dip their toe into it. Now I've seen like, you know, as we're reading all the things that are out there, I've seen some crazy things like like from like nobody, you know, successfully using them all the way to like 70% of all existing organizations are now using AI agents, which I'm, like, totally, like, BS. You know?

Daniel:

It's just not not true at all.

Chris:

Not true at all. It's a long way from the truth. But I do think a lot of organizations are kind of whiplashed right now kind of going, holy cow, what's this agent thing? I'm reading about it everywhere. And we're trying to figure out what to do.

Chris:

And that's happening at a moment of like, you know, where the like, those who are diving in and finding a use case that they can find success with, which is not easy in all cases, are are are making some big wows within within their little world. And then those who aren't are still, you know, kinda fumbling in the dark. And I think that's fair. Like, that doesn't mean that one person is smarter than the other. It just means that looking into the right use case and and getting the right people to address it and having a good business case for it makes a lot of difference.

Chris:

And, you know, as we talk about that, I think one one way of kind of leaping into that, you know, dichotomy is, as Andre Karpathy, you know, put out a post on a tweet on x. Do we still call them tweets? I don't I don't

Daniel:

I have no idea.

Chris:

I I'm not sure. But on on x, and I won't read the whole thing because it was a fairly long one. But he's basically acknowledging I mean, you're talking about this is one of the world's preeminent AI researchers that, you know, the kind like, you know, within our little AI bubble world, you know, on the technical side, he is a superstar in every possible way. And and he's kind of saying, holy cow, and I'm totally paraphrasing. He didn't actually say holy cow.

Chris:

He's kind of saying holy cow, even I at moments are am feeling a bit left behind with how fast this is changing. And in the in the context, he's kind of talking about, like, coding and stuff is that, you know, after leading in the last few years and seeing models, you know, it seems quaint to talk about models as you pointed out now. But talking about these models that are getting better and better steadily, but they still weren't doing great coding. You know, if you in in terms of that and the need for senior engineers to kinda correct it and was it more trouble to use the model and the agent to do the coding or not? And did you did it cause more did you just spend more time fixing errors?

Chris:

Well, all that really changed at the end of 2025. And, you know, you with with OPUS 4.5 and OpenAI's 5.2 model in particular, as well as several others, but those are the ones that are called out the most. Like, they got to where they could do senior level coding really well without mistakes. And and I've griped because I'm a Rust programmer that like, because that's such a small community of programmers overall that the models weren't as good as they would be in like Python and JavaScript. Well, guess what?

Chris:

It's kicking butt in Rust now. And so like, as No longer Rusty. It's no longer Rusty. And so like I, for one, as I'm as I am upskilling and as someone who has been using AI as we have gone forward in coding, like, it's now like, my workflow has changed dramatically in the last two months in terms of understanding how to effectively use coding agents to do that. And I think we're we're I don't think coding is the only area that that's impacting.

Chris:

I think there's a lot of areas where agentic AI, once you get a use case that is giving you some sense of success, is like changing the the the field that that small field that you're playing on in that. And that might be happening many times over. What do you think about

Daniel:

Yeah. Yeah. I it's interesting just to read a little bit of that tweet that you reference. Karpathy mentions, clearly, powerful alien tool was handed around, except it comes with no manual and everyone has to figure out how to hold it and operate it while the resulting magnitude nine, earthquake is rocking the profession. And he kind of ends saying, roll up your sleeves to not fall behind.

Daniel:

So yeah, I definitely have felt this Chris, just from, you know, my perspective, we get to, you know, kind of wax poetic on these episodes where it's just you and I, but from my perspective in building a company over this past year, Prediction Guard, I'm reflecting on our last board meeting and the reflection back to us as leadership was wow, like essentially making the note that product wise, all were able to advance so much more quickly without expanding your team in those last two quarters of the year than like as the company has progressed. So these are not, you know, these are obviously investors, you know, not that they have no technology, but they're not coders, but just from the output, right? The pace of development of the product and what we're able to achieve is significant. It's significant enough to be noticed in that way without the larger team that would typically have been kind of required to reach that scale or support of what supporting.

Chris:

Yeah, I wanna relay a moment that I had. And as you know, like I sent you a text over the holidays saying, Holy cow, we gotta talk about this at the beginning of the year and stuff. And I want to share with you now because you I did not relate kind of like what happened to me that made me send that text because it's very relevant to this is I had been proposing a really complicated autonomy based project for work. And I'm not going to get into specifics on what that is. But I in in, you like to take out the hype for a moment, I had spent a lot of time thinking and researching all the different things that had to go into that.

Chris:

And it was a tremendous amount of complexity involved in that and developed really complex and very detailed and specific prompt on how to get there at like a production quality where it wasn't like what we would have talked about a year ago where it was like AI slop code coming out. Mhmm. And so I finally got to this point where I tried that out. And and I like it in the matter of six minutes, it produced what I would have at least six weeks of work, at least six weeks of work in, you know, in just a matter of a a handful of minutes. Now, I had knew what I needed to put in.

Chris:

I knew a lot of that stuff and I was able to get a really good prompt going. But the actual work, like suddenly I had a large project laid out in Versus Code that had all these different things tied together. I was just I really I just literally like, I don't think I've ever had that big of an moment in coding. And that was and I was like, you gotta tell Dan about this. So that was I just wanted that was what prompted it.

Chris:

And it made me that like I flipped over and realized this is the way forward and I'm 100% in.

Daniel:

So just to pick apart a little bit of what you said, Chris, there's some highlights there that I think are takeaways from our agentic work in 2025. One of those is with, I would say no doubt at this point, these agentic workflows, especially driven by folks who have the relevant domain knowledge, are transformative in ways that are legitimately transformative, multiplicative, however you wanna say that very much. I think we can confirm that. However, I think one of the things you highlighted is some of what was highlighted throughout the year around like the MIT study of things, you know, failing Gartner says, you know, 11% of organizations have agentic AI in production and that 40% of projects will fail by 2027. I think part of that is driven maybe by two things that we've seen over this year.

Daniel:

One is you do actually need, to have a certain level of expertise to know both how to prompt and configure these, systems, but also what data sources to connect into them, how to utilize this sort of, as Karpathy puts us, this alien tool, how to hold it, how to add in, you know, an MCP server, how like, what is, like what type of automation am I really doing? How should it run? How do I integrate it in to my day to day workflow? If you have that expertise around the integration side and infusing the domain knowledge, that is a key piece of it. And without that, there can be a lot of failure.

Daniel:

Secondly, I think sometimes people are just trying to automate processes that are problematic because they're bad processes, not because the automation is bad, but they're just bad processes to begin with. So, you know, AI doesn't solve that problem.

Chris:

Yeah, there was one other takeaway that I'll throw in before we move on from this topic. That is prior to to developing the I can't call it a stub for the thing because it was too much code. But prior to that kind of final prompt that got me well into the project, I had there had been literally many hundreds of prompts before that, which got me ready for that. And I think a a key thing that I came away from that with and which I've been sharing with other people since over the last few months is that that expertise is important for how you shape prompts to get the thing you need and you learn from it. So like at no point was the AI leaving me behind.

Chris:

It would open up new doors, but I had to walk through those doors, take the learnings, and develop the next prompt from it. And I think to your point a moment ago about kind of that expertise is it took that combination of domain expertise with how you prompt your way through a long sequence of prompts to finally get to the point where like you understood the system well enough to where you could describe it in a prompt well enough so that a sophisticated agentic model could put the whole thing together in like nearly a production ready mode. So there was there was a lot of learning involved in that. So it wasn't just magic in five minutes. I was just quite taken by the by the by having gone through that long process, being able to do that final prompt and have so much produced that was at the quality level that I would have demanded it be.

Sponsors:

You know, for most developers, you've had this call. Marketing calls, sales calls, and they want a new landing page. They wanna redirect. They want designs implemented. And, course, engineering says, yeah, we'll get to it.

Sponsors:

But that bottleneck is why thousands of businesses from early stage startups to Fortune five hundreds are choosing to build their websites in Framer, where changes take minutes instead of days. So our friends at Framer, they're an enterprise grade, no code website builder that gives designers and marketers the ability to fully own your.com without having to rely on the engineering team. It works like your team's favorite design tool with real time collaboration, a robust CMS with everything you need for great SEO and advanced analytics that include integrated AB testing, changes to your Framer site go live to the web in seconds with one click publish without help from engineering. That's priceless. That keeps you on task, on target delivering features.

Sponsors:

And that's how your team reduces dependencies and reaches escape velocity. And this isn't a toy. Framer is an enterprise solution with premium hosting, enterprise grade security, and 99.99 uptime SLAs. Companies like Perplexity, Miro, and Mixpanel trust Framer for their websites, whether you want to launch a new site, test a few landing pages, or migrateyourfull.com. Framer has programs for startups, scale ups, and large enterprises to make going from an idea to live site as easy and as fast as possible.

Sponsors:

Okay. So the next step is to learn how you can get more out of your.com from a Framer specialist or get started building for free today at framer.com/practicalai for 30% off a Framer Pro annual plan. That's framer.com/practicalai for 30% off. Framer.com/practicalai. Rules and restrictions that may apply.

Daniel:

Chris, I think the other, or at least one other theme that I know that we highlighted kind of going into this year was multimodal AI. I don't think unless I'm misremembering, or not seeing the right transcript that we predicted kind of this, reasoning era with models. So like if we just look at the progression of models, which, is definitely not the whole picture as we just talked about, a lot of what happened was around agents, which, for those that are listening, maybe new, you know, parsing through these terms, an agent is not just a model. It is a model that is connected to various external systems. Some of which could be AI, of which could not be AI and actually interacts with those systems to accomplish a goal.

Daniel:

But, so we're not just talking about models anymore. We're talking about these systems, but in terms of the models, I think we predicted more multi modality kind of coming into this year, which certainly we have, right? There have been many different, you know, vision language models, video models, music models, all sorts of things, Sora, all of these, you know, things that we've seen over 2025. And then there's these other reasoning models, starting with the multimodal ones, Chris. I think at least where I'm sitting and it could be in just in my role or the types of things that I'm seeing, but in the majority, actually I think, yeah.

Daniel:

So I guess I would say all of the customer interactions that we are having and the people that I'm talking to really are using multimodal AI in terms of multimodal on the input side. I still very much, do not interact with people that are doing kind of multimodal on the output side. So what I mean by that is- That's fair. Certainly I see videos coming out of Sora as reels on social media. And so I know that that is happening, right?

Daniel:

But in terms of the business world, real business context, I definitely see, you know, video, audio, image and text going into models, but not so much coming out. Really still coming out is either text or some form of text, like some structure, like a JSON structure, a tool call, some template, some, you know, field, some whatever it is, not really, multimodal output. Maybe the exception to that might be synthesized speech, which is pretty pervasive everywhere as a thing in and of itself. So that's maybe a standout.

Chris:

I no. I I think you're right. I like, it that is a standout, but I I would have thought of that as a single mode, like, oh, on the output side. And I think you're calling out I think that there is a big opportunity here for especially when you combine it with agents in different capacities to have a richer output experience. Because at the end of the day, I mean, I know I know that, like, whereas my non AI industry family members are more taken with the videos and things like that for entertainment and but with the work that I usually do, it's more that text output.

Chris:

And I could imagine a much richer output experience to your point. It's easy to envision especially when you think about, you know, like I'm gonna dump all the different things into my input that I want it want it to process and assess for the output. But the output is still pretty basic. It might be that real time voice interaction that was so hot six months ago, you know, that everybody got into and then it kinda kinda passed and, know, we all get our expectations set very know, like, oh, okay. That's just real time voice.

Chris:

No problem. But if you were to put a bunch of things together on output where you're getting text, you're getting that real time voice in the conversational sense, you're getting supporting media, I I think it could I think it could level up. So it probably will.

Daniel:

Yeah. And I guess maybe one of the themes from this last year that we can take away as well is this rising up of the reasoning era. So reasoning models, we've talked about these on the show. Yep. Just as a reminder for people in case you missed out on our discussions for this year, in some ways, these reasoning models are mislabeled because they don't reason about anything.

Daniel:

They just produce text. What is interesting is that they produce a segment of text that imitates or mimics reasoning or a chain of thought, right? And, kind of quote unquote thinks through a problem by generating text representing that thinking through of the problem. And then they generate a final answer, which has proven to kind of help pick through more complicated tasks, maybe do more orchestration or dynamic type of workflows than what we were seeing before. And these models, I would say many of the models that I see being released now, at least in terms of that LLM flavor of models are either straight up reasoning models, which means they're always going to reason in this way.

Daniel:

They're gonna generate the reasoning tokens and then output the regular tokens. Or they are kind of conditionally, reasoning models or hybrid models that will do that some of the time and not other of the time. And, there's various implications of that. Certainly, think you see that driving certain of these agentic, you know, working towards these agentic workflows. It also, to be honest, is sometimes annoying because often like in real business applications like we're working on, if you dial in your workflow, you really don't want those reasoning tokens because they take so dang long, right?

Daniel:

You have to wait. There's so much latency introduced by waiting for these reasoning tokens that once unless you're doing this sort of very, very dynamic workflow, it's kind of annoying, that a lot of these later models have these. I would say in general, it's a good thing. So we're definitely in the reasoning era and it's been cool to see these models come about, but, you don't get anything for free. There's a lot of latency that's developed and, because these models stream output, right?

Daniel:

Every token that is generated is a inference run of the model. Meaning if you're generating thousand tokens of reasoning, that's 2,000 runs of the model that is operating on a computer with a GPU that is expensive somewhere.

Chris:

Agree with all that, but I think to some degree, it's intentionally or unintentionally and probably the former rather than the latter being driven by the organizations hosting these models. Because, know, as one example that most, you know, that everybody would know is ChatGPT, you go in and you have a choice. You know, if I'm looking at their web interface, you have a choice between kind of instant or thinking. And of course, everybody wants thinking. Do you really want instant and have thinking, you know, and then if you choose thinking, then it's just standard thinking or extended thinking.

Chris:

And so like, that plays to a human bias of like you're you're gonna go for well yeah I wanted thinking and I wanted to extend the thinking and to your point like the cost of that maybe may or may not be to you as the consumer but certainly the cost of of producing that is much more expensive with extended thinking on that, which kind of points out another thing that's happened in over the last few months that we've all heard about. And that is that we for years, we talked about the limitation of, you know, having enough GPUs being the limiting factor on moving forward. And now it is power because, you know, you can take the same GPU and use it for many inferences, but each one of those separate inferences is is taking a certain amount of power for that. So as a consumer of that, every prompt that I choose to make in a quote unquote reasoning fashion is going to be much more expensive in terms of power consumption. And we're hearing that in the news all the time these days.

Daniel:

Yeah. I guess that takes us to an interesting theme that we've seen develop around, infrastructure, hardware, energy. It's interesting to see that so much of this discussion, as you've mentioned in recent trends, and I think this will continue into 2026 and it will create some both friction and opportunity and interesting dynamics in 2026, which is this limitation and opportunity around power. Just, you know, a couple of things anecdotally, it's like, so I I went to Colorado School of Mines as my undergrad, which as the name indicates, it still has a big tie to mining and petroleum and other things. And so I have friends in the energy industry and was talking with some of them how there's now very much speculators going around and, trying to purchase and get the rights to power plants that were relatively newly constructed, but decommissioned while, you know, people were moving away from coal, speculating that, you know, these power plants will necessarily need to be turned back on.

Daniel:

Right? And, you know, other anecdotes like in our town here, Lafayette, West Lafayette, there's this huge, like, I forget number of billions of dollars investment in a chip, chip assembly plant, here, on the West Lafayette side. What's interesting to see all of the, community back and forth to get the zoning approvals and the backlash that is happening against this chip assembly plan. And I'm not saying on one side or the other of that, but what I think is interesting is you see that dynamic here, right? In China, if you want to dominate in the AI space and you need a bunch of power plants, right?

Daniel:

No city is gonna say, no, we're not going to have our power plant here. They're just going to put a power plant there, right? And so that's how this has then filtered into this geopolitical space and environment that we're in where power and AI and chip manufacturer and onshoring, all of this is what's driving now the political conversations. And so, yeah, we've seen this trend from just having access to GPUs, all the way kind of flow to these discussions around energy, infrastructure, power, which I'm sure will just continue throughout 2026.

Chris:

And it's and to delicately point at geopolitics and the implications, you know, some countries are now invading other countries and taking their oil. And, you know, that's in you know, whether which regardless of which side you're on, like, that was a notion that was kind of inconceivable. But power is is the thing that people are talking about because every nation with its drive for more and more power consumption to support not only its normal things, but but AI growth as is The United States, you see a lot of a lot of interesting things happening there. So Yeah. Yeah.

Chris:

I'll just leave that one right there.

Daniel:

I think, like, to your point, this isn't a political show. We're talking about the practicalities of AI, but I think in thinking about the trends of 2025 into 2026, you can't go into 2026 without noting that when things happen politically across the world, AI is being mentioned as a motivation for why these things are happening regardless again of, you know, who's doing right or wrong or your stance on something. We've moved from, I think at the 2024 to now the end, you know, the 2025 going into 2026, where, AI is the topic that is driving some of those policy decisions, versus I think last year, if I was to kind of summarize, were talking a lot about, well, how might AI or how might governments regulate AI as a kind of piece of their policy? Now it's almost driving the key pieces of policy in a lot of ways.

Chris:

Indeed, it's, guess going forward, it will be interesting, as we go through '26, and see how policy continues to evolve in this, because this is a level of consumption that, you know, obviously is is becoming a challenge to maintain and to even to initiate, because it's not stopping with where we're at, it's going on. So infrastructure, hardware, energy, those topics will, should be it should be a volatile year in '26 to see where things go.

Daniel:

Well, Chris, there's, of course, many, many things that that have happened in 2025. And the majority of those we've talked about so far are related to Gen AI. I think in terms of practicality, moving into 2026, we would not be practical AI. I think if we didn't highlight the fact that, you know, we recorded another episode right prior to this and I won't give away anything that's in that episode other than there was one statement that said, Hey, you know, one trend that's happening with AI models is that Gen AI models have sort of plateaued on this transformer architecture that most all of these models are based on. But predictive models still continue to advance, you know, in a quite rapid pace.

Daniel:

And what I mean by that for those listeners, that, kind of parsing through this jargon is these generative AI models, large language models, language vision models, etcetera, generate tokens or certain output like images or other things. Other models are discriminative or statistical and make predictions, of classes or forecasts or those sorts of things. And the reality is that across industry, these models still continue to provide amazing ROI and get better and better. And the tooling actually gets better around those. And actually if you, what's interesting to me, Chris, is years ago we talked about kind of this idea of AutoML, which is still a term that people use.

Daniel:

There's still some things out there related to that. This idea that we could maybe automate the parameterization of AI or statistical models, and that would kind of help us create these models better and faster. I think the reality, which is kind of interesting is everyone is talking about GenAI now, but there is actually this realization of maybe a better AutoML or maybe a better way to put it is augmented analytics or augmented ML or something like that, where actually you have these highly capable tools under the hood, whether that's SQL queries to non generative AI models, to forecasting models, to data science tools that now can actually be tied in as tools into a generative AI model that orchestrates amongst all of those and reasons over how to use those. So for example, I could have my e commerce data in a SQL database. I could have a tool that uses Facebook profit to do time series forecasting and then a generative AI model that can call those tools to pull the right data out of my SQL database, format it in a way, maybe with a generated code that's executed, send it to my time series modeling tool, which is good at time series modeling, and then gets me my forecast for 2026 for sales or something like that.

Daniel:

So actually I think it's interesting that all the discussion is really about that orchestrator model and not about these other things. Because actually it's those things that are plugged into the orchestrator that are actually creating the real multiplicative effect, the power of that system.

Chris:

I totally agree. And I think that will only get amplified you know, as you go into kind of more of a physical AI future, you know, and we've talked a lot about that in recent episodes, especially late in this past year, that as you have these orchestrators with the tooling around them, with predictive models that are now kind of enabled through agentic systems, there is so much capability out there that I don't think the public is really as aware of it. They may see drones and robots but they haven't really, in my experience, thought through what it takes for those things to come about. Yeah. Yeah.

Chris:

And so, you know, there's definitely a place for GenAI interactions that you're having with the human and in terms of how the human and the physical agent driven platform are are interacting. But kind of back to your point about predictive, you know, predictive are going up and up. And I think I think one kind of newsworthy event, kind of illustrates that is the fact that one of the what they refer to as one of the three godfathers of AI, which is, of course, Jan Lakun has left Meta, otherwise known as Facebook to people, where for about roughly a decade he was there maybe a little bit longer. But part of his tenure there were was to to draw he had kind of the academic freedom to move forward. And one of the things that he has talked about for quite some time is the fact that transformers had a limited ceiling.

Chris:

And I know we both we have, we've had those discussions lately about, you know, the limitations of Gen AI. But as he looks at the notion as along with a lot of other people in the AI industry about world models driving things forward, I think your your predictive capabilities mixed with your agentic will really drive a lot of the not only the capabilities that you just talked about with the tooling, but also things in the physical AI space. And so we may see a bit of a renaissance in those spaces going forward. As people start kind of going, I've had enough of Gen AI, it's really awesome for what it does. But I can now finally see its limitations and ceiling.

Chris:

So I'm interested in whether the upcoming year will kind of turn attention in that direction.

Daniel:

Yeah, I would say, I guess sometimes in these episodes at the beginning of the year, make predictions. I think in relation to all of what we just talked about, one of my predictions for 2026 would be that those practitioners that have the capability and knowledge to build MCP servers, to connect tools that can be orchestrated to models and to actually architect that agentic system, regardless of model. So I think like that is a wildly powerful combination. As we kind of started this conversation, we were talking about how that is part of how to get your agentic, you know, pilots and all of those things to not fail. So actually, think these, there are data scientists, software developers, etcetera, out there that are listening, now take it, I'm always wrong at predicting the future.

Daniel:

So, you know, don't trust me too, too much, but at least my own personal, you know, intuition is that focusing on, I don't even know what the term is that we'll use for this in 2026. Maybe it's AI engineer or whatever. But I think whatever that role will shape into, it will be data scientists, software developers, whoever it is, are able to come in and actually know how to spin up a system of services that are MCP servers, that are databases, that are rag systems, and then connect those things into an orchestration layer such that they can be used. I think that is shaping into a highly valuable role and something that I think will survive for some time because at least the way I would see it, those things that need connected in are so complicated across the enterprise that it's going to take a very long time for that skill of kind of integration, AI integration and tool development and tool integration to go away in any sort of meaningful way.

Chris:

I 100% agree with that. And and I think I think possibly the secret sauce on trying to put that together as a human is is I'm gonna go back and reference my little my little experience I shared in the beginning, and that is to learn how to use the tools that you have now well enough to create a workflow that allows you to leverage those tools through prompts to get all of those systems up and running. So it's not all on your shoulders as the human.

Daniel:

Yeah.

Chris:

Yeah. You're the human at the center of a great symphony of AI agents, and you have to learn to to conduct those agents in that symphony to produce way more than you could have ever done last year. And I think I think that's a doable thing, but it's a discrete skill set, and it takes a lot of flexibility and thinking and moving out of your domain of comfort to do that. So, like, be super willing to try very, very uncomfortable things. Yeah.

Chris:

So but I think that's a safe I think that's a fantastic path forward.

Daniel:

Yeah. And especially if you can drive those things to be even more sort of niche or verticalized for those out there trying to like start companies and that sort of thing. I think if there's a particular tool set within an industry that has not yet and can be tied into this level of orchestration and is necessarily complex, whether that's in manufacturing or in finance or whatever it is, and you have that domain expertise, there is definitely a window of time where it not creating a model that is able, a single model that is able to do all of that thinking, but being able to architect those tools into a system, is going to be really, really powerful. But, Chris, we're we're kind of, coming to the to the end of our discussion going into 2026. I'm wondering if if you have any thoughts on on what we'll see in 2026.

Daniel:

Are we gonna see are we gonna see quantum computing, tied in with AI? Are we going to, you know, what's gonna happen?

Chris:

So so on that one point, I don't think we're at quantum being a highly productive thing yet. I and I follow quantum a fair amount, but but that so I don't think we're quite there yet. And I think that's common people say you're, you're always ten years out or whatever that is, but we're still not there yet from seeing a fair amount of practical work on it. I'll I'll tell you what I think is is gonna change in in this coming year. And that is as we are migrating into the era of physical AI and having various types of platforms operating around us through agentic systems with lots of models both large and small participating in those.

Chris:

The cost of the the average person being able to get in there where it used to be prohibitively prohibitively expensive to do that and you had organizations, they would drive those efforts. But the maker world is really starting to see that as a possibility because GPUs and ASICs, are application specific integrated circuits and such are able to start producing AI capability going forward at a much cheaper dollar. And those are embeddable on smaller devices that you and your children will go to the store and buy, and you'll be able to implement things that just a year ago were unimaginable. They would have been far outside the family budget. And so it's no longer a commercial only interest or or an industrial or military grade interest.

Chris:

It's now something consumers have access to. And I think that as as new toys develop that that are built on this and are teaching kids that that opens up an entirely new world of capability around your house and that you'll see consumer electronics reflect this in much less expensive things instead of just having potentially a robot vacuum. You may have many little robot that are very task specific things coming into your life. And if you're not finding the thing at your local store or on online, then you just go build it yourself with your maker kits because that is becoming a real thing. It's becoming doable.

Chris:

So my prediction is we see the very beginning of the AI maker era come about at a consumer level.

Daniel:

Cool. I'm excited for it. I definitely it makes me think of I've see see all the news about CES recently. Lots of talk about robotics there, which is interesting. So my kind of set of predictions are I think a couple fold.

Daniel:

One of those which we've talked about here before, and I think is consistent with what we're seeing is, you know, models have been quite commoditized. Increases in performance of frontier models has plateaued. Open source models have essentially caught up. And so really now we're at a stage where I think like that mote of having the best model is, you know, it's not the most relevant thing. The most relevant thing is, you know, flexibility, not getting, you lock in the ability for you to use a bunch of different models, the ability for you to, you know, construct a system.

Daniel:

I think also kind of tied to that point, my second thing that I'm thinking of is just how fragmented and complicated the ecosystem is getting. I think that will carry on through 2026. We won't see as, you know, the full consolidation of that in 2026. And so I think what you'll see is all of these, so it's no longer about I'm gonna get the best model and now my company has AI and I'm set for the future. That's actually the easiest thing.

Daniel:

Like you have a model, so what? I can get one on my phone, I can get one on my laptop. Doesn't mean anything. What is problematic is if you say, okay, well, I want a system to do this. Now I need all these tools.

Daniel:

I need to connect them in a certain way. That becomes increasingly complicated. I need it to be compliant and work in a regulated industry. That becomes increasingly compliant. I need to tie in this type of data or that type of data, complexity.

Daniel:

And so you're just seeing this expansion of complexity in these AI systems, not because the models are not capable, but because the model is actually no longer the blocking point of the whole thing or the single thing in the system. And so I think if you look at something like NIST six zero one, the standard that NIST put out of how to, run secure AI, right? I did a little bit of mapping and it takes, so I tried to build up to 100% compliant with NIST six zero one and Azure AI. And by the I got up to nine different services that could get me 39% compliant with, with NIST six zero one in Azure, in Azure cloud. And so that already, you're already managing all of these different services and all of these different things in, in becomes complicated.

Daniel:

It becomes a lot of labor to do that. So I think some of the winners in this space are gonna be those that come to that complexity and tell you, hey, well, than spending up 37 different things in Azure and hiring 10 people to manage it, here's a consolidated, quick time to value way for you to get X or Y, whether that be a verticalized AI solution, a secure AI solution, whatever that might be. So those are my thoughts going into the new year.

Chris:

Excellent, excellent guidance right there. For those who are not familiar with NIST, I just wanna point out that that is a US agency called the National Institute of Standards and Technology, and they put out standards and and the 600 was was one that Dan was referring to. So if you're outside The US, can look that up. It's publicly available. But fantastic advice.

Chris:

Thank you for sharing that.

Daniel:

Yeah. And, looking forward to talking about all those things in in 2026, Chris. It's gonna be a fun year for the podcast and, new things in the works and, yeah. So thank you to our listeners for sticking with us another another year. We very much appreciate you.

Daniel:

Appreciate, sticking with us for for so long. I also appreciate the new listeners that maybe this is your your first episode that you're listening to. Welcome to the welcome to the family. Please find us on the various socials, LinkedIn, etcetera. And and, yeah, looking forward to continuing the conversation into 2026.

Chris:

It'll be a wild ride as always.

Jerod:

Alright. That's our show for this week. If you haven't checked out our website, head to practicalai.fm, and be sure to connect with us on LinkedIn, X, or Blue Sky. You'll see us posting insights related to the latest AI developments, and we would love for you to join the conversation. Thanks to our partner Prediction Guard for providing operational support for the show.

Jerod:

Check them out at predictionguard.com. Also, thanks to Breakmaster Cylinder for the beats and to you for listening. That's all for now, but you'll hear from us again next week.