Light
Dark
Product

Introducing Claude Sonnet 4.5 and the memory omni-tool in Letta

September 30, 2025

You can now use Anthropic's memory tool with Letta agents, enabling Letta agents to take full advantage of Claude Sonnet 4.5’s advanced memory capabilities to dynamically manage their own memory blocks. Agents in Letta can now not only modify existing memory blocks, but also create and delete memory blocks on-the-fly as they learn over time.

Previously, enabling an agent to create and delete memory blocks required developers to write custom tools or scripts building on the Letta API - but now, using the new built-in memory omni-tool, agents can easily handle it themselves, reorganizing their in-context memory as needed.

While Claude Sonnet 4.5 is specifically post-trained for the new memory tool and excels at using it, the tool works with any model available on Letta. And because Letta manages the underlying memory state, all memories remain model agnostic, API-accessible, and shareable between agents for collaborative learning.

Background: Agent Memory and Context Management 

Large language models face a fundamental constraint: their context windows are fixed in size and an LLM agent’s performance often degrades as the context window grows longer. This phenomenon, known as "context pollution," means that simply cramming more information into the context window isn't a viable long-term solution for agent memory. Early memory systems primarily focused on retrieval of prior conversations (RAG). MemGPT introduced the idea of agentic memory management through using tools to read/write to an external vector DB as well as rewrite segments of the agent’s context window - enabling the agent to actually learn over time.

With each new model release, the frontier models have become better and better at tool calling, making agentic memory management more powerful. Many models have also been explicitly post-trained at software engineering tasks, making them especially adept at filesystem operations - meaning that a simple filesystem can outperform purpose-built memory tools. Letta enables support for a variety of memory management tools by providing underlying storage abstractions (such as memory blocks, filesystem, and archives) that can be used to implement specialized memory tools.

Sonnet 4.5: A Model Trained for Context Management 

Memory tools in MemGPT and Letta have historically relied on the generality of LLM’s tool use capabilities. Frontier models can follow instructions from prompts to understand how to use memory tools to manage their memory and context. 

Claude 4.5 Sonnet is the first frontier model trained (publically) to be context-aware, released alongside a new memory tool for context management. Early reports suggest that the model is much more adept at understanding its context limitations and understanding how to manage memory long-term. Because the model is post-trained to understand context limitations and use a context-editing memory tool, it is able to manage its own context window more efficiently with less in-context prompting.

Using the memory tool in Letta 

Letta provides Anthropic’s `memory` tool as a new built-in tool that manages persistence of memories through memory blocks. 

Managing memory with optimized tools 

The Letta memory tool exposes the same filesystem-like API to the model as Anthropic’s version (storing memories in specific paths), but under the hood, the memory “files” are actually stored inside of Letta memory blocks. This takes advantage of Anthropic’s post-training for the tool while also managing memory storage through memory blocks, which can be visualized in the Letta Agent Development Environment (ADE).

The memory tool being called inside of Letta, which updates the human memory block stored on Letta.

Reorganizing memory with dynamic block creation 

The memory tool enables agents to dynamically restructure their memory as they learn by creating new memory blocks. An agent might start with basic memory blocks but evolve to create specialized sections for different types of information. This allows your agent to restructure its in-context memory over time. 

Example of an agent creating a memory block focused on learning how to use tools.

Manual memory management through the ADE and Letta API 

Since memories are stored as Letta memory blocks, you can:

  • Query and edit memories through the Letta API or Agent Development Environment (ADE), using the associated block_id
  • Share memory between agents by attaching existing memory blocks to new agents - enabling collaborative learning or knowledge transfer from experienced agents to new ones

This gives you direct control over agent memory without relying solely on the agent's autonomous memory management.

Next steps 

You can try out Claude Sonnet 4.5 in Letta Cloud, and use the new memory tool with any model. Although Sonnet 4.5 is specifically post-trained to the memory tool, we’ve found that other non-Anthropic models also work well with the tool.

Jul 7, 2025
Agent Memory: How to Build Agents that Learn and Remember

Traditional LLMs operate in a stateless paradigm—each interaction exists in isolation, with no knowledge carried forward from previous conversations. Agent memory solves this problem.

Jul 3, 2025
Anatomy of a Context Window: A Guide to Context Engineering

As AI agents become more sophisticated, understanding how to design and manage their context windows (via context engineering) has become crucial for developers.

May 14, 2025
Memory Blocks: The Key to Agentic Context Management

Memory blocks offer an elegant abstraction for context window management. By structuring the context into discrete, functional units, we can give LLM agents more consistent, usable memory.

Feb 13, 2025
RAG is not Agent Memory

Although RAG provides a way to connect LLMs and agents to more data than what can fit into context, traditional RAG is insufficient for building agent memory.

Feb 6, 2025
Stateful Agents: The Missing Link in LLM Intelligence

Introducing “stateful agents”: AI systems that maintain persistent memory and actually learn during deployment, not just during training.

Nov 14, 2024
The AI agents stack

Understanding the AI agents stack landscape.

Nov 7, 2024
New course on Letta with DeepLearning.AI

DeepLearning.AI has released a new course on agent memory in collaboration with Letta.

Sep 23, 2024
Announcing Letta

We are excited to publicly announce Letta.

Sep 23, 2024
MemGPT is now part of Letta

The MemGPT open source project is now part of Letta.

Jul 24, 2025
Introducing Letta Filesystem

Today we're announcing Letta Filesystem, which provides an interface for agents to organize and reference content from documents like PDFs, transcripts, documentation, and more.

Apr 17, 2025
Announcing Letta Client SDKs for Python and TypeScript

We've releasing new client SDKs (support for TypeScript and Python) and upgraded developer documentation

Apr 2, 2025
Agent File

Introducing Agent File (.af): An open file format for serializing stateful agents with persistent memory and behavior.

Jan 15, 2025
Introducing the Agent Development Environment

Introducing the Letta Agent Development Environment (ADE): Agents as Context + Tools

Dec 13, 2024
Letta v0.6.4 release

Letta v0.6.4 adds Python 3.13 support and an official TypeScript SDK.

Nov 6, 2024
Letta v0.5.2 release

Letta v0.5.2 adds tool rules, which allows you to constrain the behavior of your Letta agents similar to graphs.

Oct 23, 2024
Letta v0.5.1 release

Letta v0.5.1 adds support for auto-loading entire external tool libraries into your Letta server.

Oct 14, 2024
Letta v0.5 release

Letta v0.5 adds dynamic model (LLM) listings across multiple providers.

Oct 3, 2024
Letta v0.4.1 release

Letta v0.4.1 adds support for Composio, LangChain, and CrewAI tools.

Aug 27, 2025
Introducing Recovery-Bench: Evaluating LLMs' Ability to Recover from Mistakes

We're excited to announce Recovery-Bench, a benchmark and evaluation method for measuring how well agents can recover from errors and corrupted states.

Aug 12, 2025
Benchmarking AI Agent Memory: Is a Filesystem All You Need?

Letta Filesystem scores 74.0% of the LoCoMo benchmark by simply storing conversational histories in a file, beating out specialized memory tool libraries.

Aug 5, 2025
Building the #1 open source terminal-use agent using Letta

We built the #1 open-source agent for terminal use, achieving 42.5% overall score on Terminal-Bench ranking 4th overall and 2nd among agents using Claude 4 Sonnet.

May 29, 2025
Letta Leaderboard: Benchmarking LLMs on Agentic Memory

We're excited to announce the Letta Leaderboard, a comprehensive benchmark suite that evaluates how effectively LLMs manage agentic memory.

Apr 21, 2025
Sleep-time Compute

Sleep-time compute is a new way to scale AI capabilities: letting models "think" during downtime. Instead of sitting idle between tasks, AI agents can now use their "sleep" time to process information and form new connections by rewriting their memory state.