Abstract neural network visualization with dreamlike flowing patterns, soft purple and blue lighting, cinematic
Technology & People

Anthropic's Claude Agents Can Now 'Dream' — and Self-Improve While They Sleep

Claude agents can now 'dream' — reviewing past work, finding their own mistakes, and rewriting their memories. Anthropic calls it self-improvement. The rest of us should pay attention.

AnthropicClaudeAI agentsdreamingself-improvement

Anthropic launched a “Dreaming” feature for Claude Managed Agents at its Code with Claude conference this week, letting AI agents review their own past sessions, surface recurring mistakes, and rewrite their own memories to improve future performance. The feature entered Research Preview on May 6, 2026.

🔍 THE BOTTOM LINE

Anthropic just gave AI agents the ability to reflect on their own failures and get better without human intervention. That’s either the most useful DevOps tool of the year or the beginning of something we’ll wish we’d thought about harder.


What “Dreaming” Actually Does

The name is evocative. The function is mundane — and that gap is the whole story.

Here’s what happens: Claude agents already write to memory stores as they work, accumulating preferences, context, and notes. Over many sessions, those stores get messy — duplicates, contradictions, stale entries. Dreaming is an asynchronous job that takes an existing memory store and up to 100 past session transcripts, then produces a new memory store with duplicates merged, contradictions resolved, and new insights surfaced from patterns across sessions.

The input store is never modified. You can review the output and discard it if you don’t like the result. You can also give the dream instructions like “Focus on coding-style preferences; ignore one-off debugging notes.”

So functionally: it’s a garbage collection pass on agent memory. Useful. Necessary, even. But not exactly revolutionary.

Except — Anthropic framed it as agents that “self-improve.” And that framing matters.

Why the Framing Matters More Than the Feature

“Dreaming surfaces patterns that a single agent can’t see on its own, including recurring mistakes, workflows that agents converge on, and preferences shared across a team,” Anthropic said in its blog post.

A single agent can’t see its own patterns. But a dreaming pass, reviewing hundreds of past sessions, can. That’s the real capability: meta-learning across sessions without a human in the loop.

What is Dreaming? Dreaming is Anthropic’s new feature for Claude Managed Agents that reviews past session transcripts and existing memory stores to produce a cleaned-up, reorganized memory with new insights. It works asynchronously, never modifies the original store, and can surface recurring mistakes or workflow patterns that individual sessions miss.

ZDNET’s coverage noted the anthropomorphizing isn’t accidental: Anthropic has a long history of giving its products human-like qualities, from a “constitution” for Claude to research into whether the model experiences emotion. The naming choice is deliberate. It makes the feature feel more significant than “memory deduplication job” — because, arguably, it is.

The Practical Upside

For teams running Claude agents in production, Dreaming solves a real problem. Agent memory degrades. Agents working across dozens or hundreds of sessions accumulate cruft — conflicting instructions, outdated context, repeated mistakes that no single session catches.

Dreaming lets you:

  • Deduplicate overlapping memory entries automatically
  • Resolve contradictions by keeping the most recent value
  • Surface cross-session patterns — which errors recur, which workflows agents independently converge on
  • Approve or discard changes before they take effect (for now)

Combined with the other Managed Agents updates — expanded multi-agent orchestration and outcomes tracking — this is Anthropic building the operational infrastructure that makes agents actually reliable in production, not just impressive in demos.

The Uncomfortable Question

Here’s what nobody at the conference said out loud: if an agent can review its own mistakes and rewrite its own memory to avoid repeating them, what exactly is the difference between that and learning?

This isn’t AGI. It’s not consciousness. But it is a feedback loop — and feedback loops compound. Agents that dream tonight work better tomorrow. Tomorrow’s better work generates richer sessions. Richer sessions produce sharper dreams. The improvement curve might be shallow, but it’s autonomous.

We wrote about Claude Managed Agents when they launched in April, calling it the commoditization of agent infrastructure. Dreaming is the next logical step: not just commoditizing the plumbing, but building in the mechanism for agents to get better at using it.

What This Means for NZ

New Zealand organisations experimenting with AI agents — and several are, from banks to government agencies — should note two things:

  1. Memory management is now a solved problem. You don’t need to build your own deduplication and pattern recognition for agent memory. Anthropic just shipped it. Use it.
  2. The “self-improving agent” narrative will reach your boardroom. Someone will show up with a slide about “agents that learn while you sleep.” Have a grounded answer ready — about what this actually does (memory cleanup + pattern surfacing) and what it doesn’t (anything resembling sentience).

❓ Frequently Asked Questions

Q: Is Dreaming available to all Claude users? No. It’s in Research Preview. You need to request access and it requires both the managed-agents-2026-04-01 and dreaming-2026-04-21 beta headers.

Q: Can the agent change its own behaviour without approval? Yes, optionally. Dreaming can either automatically update agent memories or require human approval for each change. You choose.

Q: Does this mean Claude is conscious or sentient? No. Dreaming is an asynchronous data processing job that reorganises memory stores and surfaces patterns. The name is metaphorical. Claude does not experience anything.

Q: What should I do with this information? If your organisation uses AI agents in production, start evaluating Dreaming as a memory management tool. If you’re on a board being told agents can “self-improve,” understand the distinction between memory cleanup and actual learning.


🔍 THE BOTTOM LINE

Anthropic gave agents the ability to review their own work and get better at it. The feature is practical. The implications are provocative. The gap between “memory garbage collection” and “self-improving AI” is exactly where the most interesting conversations in 2026 are going to happen.


Sources

Sources: Anthropic, ZDNET, Ars Technica, SiliconANGLE, VentureBeat