Episode 14
Agentic Memory: Stateful AI & RAG Extensions (Chapter 16)
Discover how agentic memory is transforming AI from forgetful assistants into adaptive, stateful partners that remember, learn, and evolve over time. In this episode, we unpack Chapter 16 of Keith Bourne’s 'Unlocking Data with Generative AI and RAG' and explore the strategic impact of extending Retrieval-Augmented Generation (RAG) with dynamic memory systems designed for real-world business advantage.
In this episode:
- What agentic memory is and why it matters for AI-driven products and services
- Comparison of leading agentic memory tools: Mem0, LangMem, Zep, and Graphiti
- How different memory types (working, episodic, semantic, procedural) enable smarter AI agents
- Real-world use cases across finance, healthcare, education, and tech support
- Technical architecture insights and key trade-offs for leadership decisions
- Challenges around memory maintenance, privacy, and compliance
Key tools & technologies mentioned:
- Mem0
- LangMem
- Zep
- Graphiti
- Vector databases
- Knowledge graphs
Timestamps:
0:00 - Introduction to Agentic Memory & RAG
3:30 - The strategic shift: from forgetful bots to adaptive AI partners
6:00 - Why now? Advances enabling stateful AI
8:30 - The CoALA framework: modeling AI memory like human cognition
11:00 - Tool head-to-head: Mem0, LangMem, Zep/Graphiti
14:00 - Under the hood: memory extraction and storage techniques
16:00 - Business impact: accuracy, latency, ROI
17:30 - Reality check: challenges and risks
19:00 - Real-world applications & leadership takeaways
Resources:
- "Unlocking Data with Generative AI and RAG" by Keith Bourne - Search for 'Keith Bourne' on Amazon and grab the 2nd edition
- Memriq AI - https://memriq.ai
Transcript
MEMRIQ INFERENCE DIGEST - LEADERSHIP EDITION Episode: Agentic Memory: Chapter 16 Deep Dive into Stateful AI & RAG Extensions
MORGAN:Welcome to Memriq Inference Digest - Leadership Edition, your go-to podcast for cutting-edge AI insights tailored for decision makers. This show is brought to you by Memriq AI, a content studio building tools and resources for AI practitioners — check them out at Memriq.ai.
CASEY:Today, we’re diving into a fascinating topic from Chapter 16 of 'Unlocking Data with Generative AI and RAG' by Keith Bourne: Agentic Memory — Extending Retrieval-Augmented Generation, or RAG, with Stateful Intelligence. We’ll explore how AI agents can go beyond simple one-time responses and actually remember, learn, and adapt over time to become more like trusted colleagues.
MORGAN:If you want to go deeper beyond our highlights — with detailed diagrams, thorough explanations, and hands-on code labs — just search for Keith Bourne on Amazon and grab the second edition of his book.
CASEY:And we’re thrilled to have Keith himself joining us throughout the episode to share insider insights, behind-the-scenes thinking, and real-world experience that didn’t all make it into the book.
MORGAN:Over the next 20 minutes, we’ll unpack the strategic implications of agentic memory, compare some leading tools like Mem0, LangMem, Zep and Graphiti, discuss real business use cases, and even debate which approach suits which scenario best. So, let’s get started.
JORDAN:Here’s something that really flips the script on how most people think about AI chatbots. What if I told you that agentic memory transforms AI from a forgetful, one-off responder into a smart, adaptive partner that actually learns and remembers your preferences over time?
MORGAN:Wait, so these aren’t just chatbots that start fresh each time?
JORDAN:Exactly. Unlike traditional bots that forget everything once the conversation ends, agents with memory build ongoing relationships by recalling past interactions and evolving their knowledge. It’s like moving from a customer service rep who answers a question once, to a colleague who knows your history, context, and even anticipates your needs.
CASEY:That’s a powerful shift. It means AI can act more like a knowledgeable teammate rather than a transactional assistant.
MORGAN:And I imagine that kind of personalization could massively improve customer engagement and operational efficiency.
JORDAN:Spot on. It opens up strategic advantages for companies willing to invest in AI that remembers — driving loyalty, reducing friction, and enabling smarter decision-making over time.
CASEY:To boil it down in one sentence: Agentic memory extends Retrieval-Augmented Generation by equipping AI agents with memory systems that let them remember, learn, and adapt through structured, evolving knowledge stores.
MORGAN:And the tools leading the pack here are Mem0, LangMem, Zep, and Graphiti — each taking a different approach to how that memory is stored and used.
CASEY:So if you remember nothing else, think of agentic memory as the technology turning AI from a flash-in-the-pan answer machine into a strategic, long-term partner in your business.
JORDAN:So, why is now the moment for agentic memory? The problem before was that most AI models had a limited “context window”—basically, they could only “remember” a small chunk of recent conversation or data. That meant if you wanted AI to recall anything beyond that, it simply forgot.
CASEY:Right, and that short context window was a real bottleneck for applications needing ongoing, personalized relationships, like in finance, healthcare, or education.
JORDAN:Exactly. Now, advances in memory architectures and the ability to dramatically expand these context windows have changed the game. AI can maintain state—what we call “stateful AI”—which means it keeps track of history over long periods, not just minutes.
MORGAN:Who’s actually adopting these systems now?
JORDAN:Leading financial advisors are using agentic memory to remember client portfolios and risk profiles over months. Healthcare assistants track patient histories and preferences longitudinally. Educational platforms adapt to individual students’ progress. Basically, companies focused on customer experience and operational continuity see massive value here.
CASEY:So the convergence of business demand and technical readiness creates a unique window for investment in stateful AI.
JORDAN:Spot on — this isn’t just a cool tech trick. It’s a strategic lever to unlock competitive advantage today.
TAYLOR:Let’s get to the heart of it. Agentic memory equips AI agents with multiple types of memory — think of it like how humans remember different things in different ways. We have working memory for short-term context, episodic memory for experiences, semantic memory for facts, and procedural memory for skills and processes.
MORGAN:So it’s not just one big “memory bank” but a more nuanced system?
TAYLOR:Exactly. This is a core idea in the CoALA framework that Keith introduces in the book — it models AI memory architecture after human cognition to create richer, more adaptive interactions. Traditional RAG systems focus on retrieving documents or facts statically — a bit like referencing a library on demand. Agentic memory extends this by creating dynamic, personalized knowledge stores that evolve. So the AI doesn’t just pull information — it learns from prior interactions to generate smarter responses.
MORGAN:Keith, as the author, what made you decide to cover agentic memory so prominently in your book?
KEITH:Great question, Morgan. When I was consulting with companies deploying AI at scale, I saw a clear gap between the promise of RAG and the reality: AI that forgets too quickly can’t build trust or provide lasting value. Agentic memory is the bridge — it transforms AI into a partner that truly understands context over time. Covering it early helps leaders grasp the strategic shift from static retrieval to dynamic intelligence.
TAYLOR:That really frames the architectural decisions companies need to make upfront.
TAYLOR:Now, let’s compare how some leading tools approach agentic memory, each with their own trade-offs.
CASEY:I’ll push back when needed — these sound great, but what’s the catch in each?
TAYLOR:First up, Mem0. It offers a unified long-term memory store that prioritizes simplicity and performance, without explicitly separating memory types. That’s great if you want fast, easy deployment with solid accuracy.
CASEY:But does that mean it might miss nuances by treating all memories the same?
TAYLOR:Possibly. Then there’s LangMem, which explicitly separates memory into working, episodic, semantic, and procedural types — aligning with the CoALA framework. This allows for more nuanced memory management and supports procedural memory, which means the AI can learn skills or routines over time.
MORGAN:Sounds more sophisticated but probably more complex to build.
TAYLOR:Exactly. Next, Zep paired with Graphiti uses temporal knowledge graphs to track how memories evolve over time, which aids historical reasoning and auditability. However, it doesn’t natively support procedural memory.
CASEY:So great for tracking and explaining decisions over time but limited in adaptive behavior?
TAYLOR:Right. Each has strengths and trade-offs around complexity, capability, and operational needs.
CASEY:So from a leadership perspective, you’d choose Mem0 when you want speed and simplicity, LangMem when you need detailed cognitive memory modeling, and Zep/Graphiti when audit trails and temporal reasoning are critical.
TAYLOR:Perfect summary.
ALEX:Let’s peel back the curtain and walk through how agentic memory works under the hood — the step-by-step flow. First, the AI captures raw interactions — conversations, user inputs, or sensor data. This goes into a memory extraction pipeline, which distills the information into structured memories. These memories are categorized into four types: Working memory: Think of this as the AI’s short-term scratchpad to maintain immediate context. Episodic memory: Stores specific experiences like “the customer asked about product X last month.” Semantic memory: Facts and general knowledge—like “product X costs $100.” Procedural memory: Skills and routines, such as “how to troubleshoot a common issue.”
ALEX:These memories are stored using different technologies. Vector databases enable efficient semantic search by converting memories into numerical “embeddings” — basically, a way for the AI to find relevant info based on meaning rather than keywords.
MORGAN:So it’s like the AI’s brain translating memories into a language it can search quickly?
ALEX:Exactly. Other systems use knowledge graphs to map relationships and temporal metadata to track how memories change over time. Mem0, for instance, focuses on a unified vector store optimized for performance — it uses clever indexing and memory curation strategies like deduplication and summarization to keep things lean. LangMem separates memory types and implements procedural memory to allow agents to learn new behaviors, which is a big step towards more autonomous AI. Zep with Graphiti adds a temporal dimension — think of it as the AI’s diary, noting not just facts but when and how they changed.
ALEX:Keith, your book has extensive code labs on these mechanics — what’s the one thing you want readers to really internalize here?
KEITH:It’s that agentic memory isn’t just about storing data; it’s about designing how AI thinks over time. Understanding these different memory types and how they interact is key to building agents that adapt meaningfully rather than just regurgitating static info. The labs walk through how to implement these concepts practically, which really helps internalize the architecture.
ALEX:That perspective is vital. It’s about building AI that learns like we do — not just recalling but evolving.
ALEX:Let’s talk numbers, because what really matters to leadership is the impact. Mem0 boasts a 26% accuracy improvement over standard OpenAI memory approaches, with a whopping 90% reduction in latency and token consumption. That’s huge — faster responses, lower costs, and better answers.
MORGAN:Wow, 90% reduction in latency? That’s a game changer for real-time applications.
ALEX:Absolutely. Zep with Graphiti hits 94.8% accuracy on deep memory retrieval tests and improves long-memory evaluation by 18.5%, while also cutting latency by 90%.
CASEY:So these aren’t just incremental gains — they’re leaps forward in both performance and efficiency.
ALEX:Exactly. The practical effect? More coherent conversations, better personalization, and AI that actually learns from history instead of starting fresh each time.
MORGAN:That translates directly into better customer experiences and operational efficiency — a clear ROI.
ALEX:Right, but it’s also worth noting these gains come with complexity and the need for ongoing maintenance, which we’ll touch on next.
CASEY:Speaking of which, let’s get real. What can go wrong with agentic memory?
MORGAN:I’m bracing for some tough truths here.
CASEY:Implementing these systems is complex — you need high-quality data capture pipelines and constant maintenance to prevent “memory bloat,” where irrelevant or outdated memories clutter the system, degrading performance.
JORDAN:Memory sclerosis is another issue — where memories become stale or biased over time, leading to poor decisions.
CASEY:Procedural memory, which is key for adaptive behavior, is still less commonly supported and can be tricky to get right.
MORGAN:What about privacy? Storing personal memories sounds like a regulatory minefield.
CASEY:Exactly. Data isolation, GDPR compliance, and user consent are critical challenges that can’t be overlooked.
MORGAN:Keith, from your consulting experience, what’s the biggest mistake organizations make with agentic memory?
KEITH:The biggest pitfall is treating it as a “set and forget” system. These memory architectures require ongoing curation — pruning irrelevant memories, updating models, and balancing privacy with personalization. Neglect this, and you risk degraded performance and compliance issues. Also, rushing implementation without clear business goals leads to wasted effort.
CASEY:So thoughtful strategy and continuous investment are non-negotiable.
KEITH:Absolutely. The RAG book is honest about these limitations because we want leaders to be fully prepared.
SAM:Let’s bring this home with some real-world examples. In finance, some advisory firms use agentic memory to track client portfolios and risk tolerance over years. This enables advisors to deliver personalized recommendations without making clients repeat themselves every time.
MORGAN:That’s a real customer loyalty win.
SAM:In healthcare, virtual assistants remember patient symptoms, treatments, and preferences longitudinally, helping clinicians get a fuller picture without sifting through piles of records.
CASEY:That must improve care continuity and reduce errors.
SAM:Exactly. Tech support bots build memory from past debugging conversations, so they get smarter at troubleshooting and reduce repeat escalations.
JORDAN:And in education, AI tutors adapt to students’ learning styles and progress over time, personalizing lesson plans dynamically.
SAM:Across industries, agentic memory is a key differentiator for companies that want AI to truly become a strategic partner rather than just a tool.
SAM:Now, let’s stir up some debate. Imagine you’re building an AI customer support agent for a regulated financial services firm that needs audit trails and evolving compliance rules.
MORGAN:I’d lean into Zep with Graphiti. The temporal knowledge graph’s ability to track changes over time and provide auditability is a must-have in finance.
CASEY:But what about Mem0? Its simplicity and performance could mean faster deployments and lower operational overhead.
TAYLOR:True, but Mem0 doesn’t natively support procedural memory or temporal reasoning, which could limit adaptability in a complex domain like finance.
ALEX:LangMem’s explicit memory types and procedural capabilities might offer a middle ground — enabling the agent to learn compliance routines while keeping memory structured.
SAM:So it boils down to a trade-off between auditability and complexity. Zep/Graphiti for compliance-heavy needs; Mem0 for rapid, unified memory; LangMem when procedural learning is critical.
MORGAN:Could a hybrid approach work?
TAYLOR:Absolutely. Combining Zep’s temporal modeling with LangMem’s procedural memory could cover all bases — but that comes with integration complexity.
SAM:Leaders need to weigh these trade-offs carefully based on product goals, regulatory needs, and team capabilities.
SAM:Before we wrap, some quick tips for leaders considering agentic memory. Start with a clear memory strategy—define what types of memory your AI really needs based on use case.
MORGAN:Avoid the temptation to store everything—implement memory curation like deduplication and decay to keep the system efficient.
CASEY:Use semantic similarity search with vector databases for fast, meaningful retrieval — it’s way smarter than keyword matching.
JORDAN:Balance personal and community memory carefully: personal memories drive customization, while community memories capture shared knowledge.
ALEX:And build robust memory extraction pipelines that translate raw data into structured memories—this often gets overlooked but is foundational.
SAM:These patterns help ensure your agentic memory system scales and delivers value without becoming a maintenance nightmare.
MORGAN:Just a quick reminder: we’re giving you the highlights here, but Keith’s book goes much deeper — with detailed diagrams, thorough explanations, and hands-on code labs that let you build these systems yourself. Search for Keith Bourne on Amazon and grab the second edition of Unlocking Data with Generative AI and RAG.
MORGAN:This podcast is produced by Memriq AI — an AI consultancy and content studio building tools and resources for AI practitioners.
CASEY:Memriq helps engineers and leaders stay current with the rapidly evolving AI landscape. Head to Memriq.ai for more deep-dives, practical guides, and cutting-edge research breakdowns.
SAM:Looking ahead, there are still some big challenges. Scaling memory systems without slowing down as data grows exponentially is a major engineering hurdle.
TAYLOR:Developing adaptive procedural memory that reliably evolves agent behavior remains an open frontier.
CASEY:And the tension between personalization and privacy compliance is only going to get tougher, especially with new regulations emerging.
ALEX:Plus, we need better evaluation metrics to truly measure long-term learning and behavioral adaptation — current tests often miss the bigger picture.
SAM:Leaders should watch these areas closely and plan investments accordingly to stay ahead.
MORGAN:My key takeaway? Agentic memory is less a feature and more a strategic transformation — AI that remembers is AI that builds trust.
CASEY:For me, the biggest lesson is that complexity is real here — but with clear focus and ongoing curation, the payoff can be enormous.
JORDAN:I’m excited by how agentic memory brings AI closer to human-like relationships, unlocking richer, more personal experiences.
TAYLOR:The CoALA framework is a game-changer — it gives leaders a clear mental model to evaluate memory architectures.
ALEX:Remember, it’s not just about storage — it’s about designing how AI thinks and learns over time.
SAM:Real-world deployments show this isn’t theoretical — agentic memory is driving differentiation right now.
KEITH:As the author, the one thing I hope you take away is that integrating agentic memory thoughtfully isn’t just a technical challenge — it’s a leadership imperative to unlock AI’s true potential for your business.
MORGAN:Keith, thanks for giving us the inside scoop today.
KEITH:My pleasure — I hope this inspires you to dig into the book and build something amazing with agentic memory.
CASEY:And thanks to everyone listening. Remember, we covered key concepts today, but Keith’s book goes much deeper — detailed diagrams, thorough explanations, and hands-on code labs that let you build this stuff yourself. Search for Keith Bourne on Amazon and grab the second edition of Unlocking Data with Generative AI and RAG.
MORGAN:Thanks for tuning in to Memriq Inference Digest - Leadership Edition. Until next time!
