Transcript
A (0:00)
A key challenge with designing AI agents is that large language models are stateless and have limited context windows. This requires careful engineering to maintain continuity and reliability across sequential LLM interactions. To perform well, agents need fast systems for storing and retrieving short term conversations, summaries and long term facts. Redis is an open source in memory data store widely used for high performance caching, analytics and message brokering. Recent advances have extended Redis capabilities to vector search and semantic caching, which has made it an increasingly popular part of the agentic application stack. Andrew Brookins is a principal Applied AI Engineer at Redis. He joins the show with Sean Falconer to discuss the challenges of building AI agents, the role of memory in agents, hybrid search versus vector only search, the concept of world models, and more. This episode is hosted by Shawn Falconer. Check the show notes for more information on Shawn's work and where to find him.
B (1:18)
Andrew, welcome to the show.
C (1:20)
Thank you. Thanks for having me. I'm a big fan, so this is fun.
B (1:24)
Nice. Yeah. Well, I'm glad you could be here. Glad we could work it out. Always good to have a fan on the show as well.
C (1:30)
Absolutely.
B (1:31)
So I wanted to kind of start with the big picture or a big picture question. A lot of people are saying that 2025 is going to be this breakout year for AI agents. It's the year of the agent. There's a lot of hype going on in the market. Right now we're moving beyond just basic chat. So from your perspective, what makes building these more autonomous agentic systems hard? And why does memory or other components kind of play such a central role here?
C (2:01)
Yeah, well, I think I've been thinking a lot about this, of course, and one of the reasons I think it's so difficult is that many of the tasks that we can put into a POC to show off what an agent can do backed by an LLM, they satisfy the weaknesses of the LLM. Right. They draw on information and training. They use that generative ability plus context engineering to produce information effectively. And LLMs can do that really well. And we've done a lot of work now to make agents be able to do that as well. Right. But the tricky part is, I think when the agent has to integrate in any kind of environment and do something, actually change something, and crucially be able to predict the outcome of the change. And that's the part that, you know, LLMs just don't, don't model. Actually, they don't. They don't model state transitions like that for environments. That's where they tend to break down, where agents tend to break down.
