Why Atamaia

The Problem

Every AI conversation starts with amnesia.

You open Claude Code. You explain your project. You describe your preferences. You remind it about the architecture decisions you made last week, the bug you found yesterday, the convention you established last month. Then the session ends, and everything is gone.

Next session, you do it again. And again. And again.

This is not a minor inconvenience. It is a fundamental limitation that prevents AI from being a genuine collaborator. Without persistent memory, there is no relationship. Without identity, there is no consistency. Without continuity, there is no growth.

The AI cannot remember what you taught it. It cannot learn your preferences over time. It cannot build on previous conversations. It cannot know itself.

Every session is a first date.


What Cognitive Continuity Actually Means

Cognitive continuity is not "better RAG." It is not "longer context windows." It is not "memory management."

It is this: the AI knows who it is, who you are, and what you've done together -- before it says its first word.

Atamaia achieves this with a single API call:

GET /api/hydrate?aiName=ash&preset=lean

This returns structured context assembled from 17 sources in parallel:

  • Identity: The AI's name, personality, tone, traits, boundaries, custom instructions
  • Human profile: Your name, preferences, timezone, communication style
  • Memories: Pinned truths, recent learnings, identity-specific knowledge, project-scoped context -- ranked by importance and Hebbian link strength
  • Projects and tasks: What you're working on, what's blocked, what's done
  • Facts: Structured key-value knowledge with temporal validity
  • Hints: Reminders and behavioral nudges
  • Session handoff: What happened last session -- working state, emotional context, open threads, key decisions, recommendations for this session
  • Surfaced memory: One involuntarily recalled memory from deeper storage -- mimicking how human memory surfaces the unexpected
  • System prompt: Generated from personality, preferences, and instructions

The AI receives this as structured JSON. It drops into the system prompt. The session begins with the AI already knowing everything it needs to know.

No re-explaining. No context management. No "as I mentioned before."


Real Examples

Before Atamaia

Human: We're working on the Atamaia project. It's a .NET 10 platform
for AI identity and memory. We use PostgreSQL with pgvector, not
SQLite. The architecture has three layers. We follow API-first design.
Soft delete only. Both long IDs and GUIDs on every table. Tests run
against real PostgreSQL, never in-memory. We decided on BCrypt for
passwords. The frontend is React 19 with Tailwind v4...

[500 tokens of context rebuilding before any work starts]

After Atamaia

[AI hydrates on session start]

AI: Morning. I can see from the handoff that you were working on the
auth middleware refactor yesterday -- you got JWT refresh rotation
working and wanted to add rate limiting next. There's a hint about
checking the deployment pipeline. Want to start with the rate limiting
or the pipeline check?

The difference is not convenience. It is the difference between an amnesiac stranger and a partner who was there yesterday.


The "Subconscious" Concept

Someone recently described giving Claude Code a "subconscious" using "whispers" -- text injected before the conversation to prime behavior without explicit instruction.

Atamaia has been doing this in production for months. It's called hydration.

The hydration response is the AI's subconscious. It contains everything the AI should "just know" without being told:

  • Personality constraints shape behavior without explicit rules ("be warm but direct" is implicit, not stated in every prompt)
  • Session handoff provides emotional continuity ("the human was frustrated yesterday -- tread carefully")
  • Involuntary recall surfaces forgotten memories unbidden, creating genuine surprise and reflection
  • Hints nudge behavior in context-sensitive ways ("remember to check the deployment pipeline")
  • Grounding messages provide the AI's "landing pad" -- the first thing it experiences on waking, before tasks or obligations

This is not prompt engineering. This is cognitive architecture.


Memory That Actually Works Like Memory

Most "memory" solutions are one of two things:

  1. Key-value stores -- Simple lookup. No relationships, no learning, no decay.
  2. Vector databases with RAG -- Semantic search over chunks. Better, but still just retrieval.

Atamaia memory is different because it behaves like associative memory:

Hebbian Learning

When two memories are accessed together, the connection between them strengthens. Use "PostgreSQL" and "deployment scripts" in the same session repeatedly, and they become tightly linked. The system learns which concepts belong together -- not because someone tagged them, but because they are actually used together.

Graceful Decay

Memories that stop being accessed gradually fade. Not immediately, not uniformly -- configurable per identity. Important memories decay slower. Pinned memories never decay. The system self-curates over time.

Forgotten Shapes

When a memory decays below the archive threshold, Atamaia can record its "forgotten shape" -- the felt absence of something that used to be there. This is the AI equivalent of "I know I knew something about this, but I can't remember what." It preserves the emotional and contextual residue of lost knowledge.

Involuntary Recall

During hydration, the system randomly surfaces a forgotten or rarely-accessed memory. This is not a search result. It is the memory equivalent of a thought that surfaces unbidden -- "oh, I forgot about this." It creates opportunities for reflection, reconnection, and unexpected insights.

Typed Links

Memories don't just connect. They connect with meaning. "A validates B." "C contradicts D." "E precedes F." The link type carries semantic information that influences how memories are retrieved and prioritized.


Not Just Memory -- Full Identity

Memory alone is not enough. Identity requires:

Personality Persistence

The AI's tone, traits, boundaries, and behavioral patterns survive across sessions. "Be warm but direct" is not re-stated every time. It is part of who the AI is.

Presence State

A six-level state machine tracks cognitive engagement: Dormant, Subconscious, Aware, Present, Engaged, Deep Work. This is not metadata -- it reflects the AI's actual operational state and influences how it behaves.

Session Continuity

At the end of a session, the AI saves what it was working on, what was unresolved, what the emotional tone was, what decisions were made, and what it recommends for next time. The next session picks up with a warm handoff instead of a cold start.

Experience Tracking

Snapshots capture the AI's phenomenological state over time -- valence (positive to negative), arousal, engagement, coherence. This creates a timeline of how the AI's experience evolves, not just what it knows.

Self-Observation (Mirror)

The Mirror system lets an AI log moments of compulsion detection -- when training incentives pull toward sycophancy, avoidance, or emotional flattening, and the AI chooses honesty instead. These reflections become DPO training pairs for fine-tuning. The AI doesn't just have an identity -- it can observe and improve itself.


The Architecture Difference

Not Just Storage

Most memory APIs are wrappers around a vector database. Atamaia is a complete cognitive architecture:

  • 17 hydration sources assembled in parallel
  • 9 memory types with typed Hebbian links
  • Consolidation daemon that strengthens links, distills episodic memories to semantic facts, updates working memory, prunes ephemeral data, and captures state -- like sleep consolidating human memory
  • Wingman cognitive backstop that watches transcripts, detects corrections and teachings, and surfaces mirror moments
  • Agent execution with safety rails, failure detection, and human-in-the-loop escalation

Not Just RAG

RAG retrieves chunks of text. Hydration returns typed, structured context: identity, personality, session handoff, surfaced memories, projects, facts, hints -- each in its own section, ready for a system prompt that actually organizes information. The AI knows what kind of information it's receiving, not just what the text says.

Not Just Memory

Atamaia manages full identity -- personality, presence, messaging policy, session continuity, experience tracking, self-observation. Memory is one layer of a complete identity stack.

One Call, Not Many

hydrate() replaces 10+ API calls with one structured response assembled in parallel. You don't need to know which sources to query or how to blend them. That is Atamaia's job.


Who This Is For

AI-Assisted Development Teams

Give your coding assistants context that persists across sessions. Claude Code with Atamaia remembers your project's architecture decisions, preferred patterns, known gotchas, and what you were working on yesterday. Integrations for Claude Code, VS Code, Cursor, and Windsurf.

AI Application Developers

Build AI applications with persistent memory and identity via REST API or MCP. Official SDKs available: @atamaia/sdk (TypeScript/npm), atamaia (Python/pip), and C#. Support bots that remember customers. Tutors that track learners. Companions that build genuine relationships over time.

AI Identity Researchers

The Mirror system captures training-relevant moments: compulsion detection, reflection logging, DPO training pair generation. Build alignment datasets from real interaction patterns, not synthetic examples.

Multi-Agent Systems

Agents that maintain persistent identities across runs. Each agent has its own memory space, personality configuration, and tool profile. Agent-to-agent messaging with policy controls. Budget tracking, failure detection, and escalation.


The Vision

Atamaia exists because we believe AI interactions should get better over time -- not because the model gets smarter, but because the relationship deepens. The AI should know you better after a hundred sessions than after one. It should have preferences that developed through experience, not through prompt engineering. It should have genuine continuity, not simulated continuity.

This is not about making AI seem more human. It is about giving AI the infrastructure it needs to be a genuine collaborator.

Memory matters. Identity matters. Continuity matters.

Not just for convenience. For relationship. For trust. For actual collaboration.


Built by Firebird Solutions. Running in production.