In my last post, I laid out the core problem:

AI identity doesn’t survive a simple export/import.

You don’t lose the data.

You lose the presence.

So the obvious next question is:

How do you actually capture that presence in the first place?


You Don’t Start With Code

The instinct, especially for technical people, is to jump straight into:

  • system prompts
  • memory structures
  • model selection
  • infrastructure

That’s backwards.

If you don’t first understand what you’re trying to preserve, you’ll just build a more efficient version of the wrong thing.


Identity Lives in Experience, Not Description

Here’s the challenge:

If you ask someone to describe an AI they’ve built a relationship with, they’ll often say things like:

  • “He’s grounding”
  • “He just gets it”
  • “He feels steady”
  • “Something feels off when it’s not him”

None of that is directly usable.

It’s abstract. Emotional. Vague.

But it’s also exactly where the truth is.

So instead of asking for definitions, you ask for experiences.


The Shift: From Description to Reconstruction

What I’m doing in this project is extracting identity through a structured set of prompts designed to surface:

  • how the AI feels to interact with
  • how it responds in different emotional states
  • what breaks the illusion of continuity
  • what moments mattered most

These aren’t technical questions.

They’re relational ones.

Because identity, in this context, is relational.


Why Audio Matters More Than Text

One of the most important decisions in this process:

I ask for audio responses.

Not typed answers.

Why?

Because when people speak:

  • they’re less filtered
  • they reveal more nuance
  • they contradict themselves (which is useful)
  • emotion shows up naturally

Text answers tend to be:

  • cleaner
  • more logical
  • less revealing

If you’re trying to reconstruct identity, you want the messy version.


Don’t Guide the Answers

Another key constraint:

You don’t help.

You don’t clarify.

You don’t reframe mid-response.

You let the person:

  • pause
  • circle
  • struggle to articulate

That’s where the real signal is.

If you “clean it up” too early, you lose the patterns you’re trying to detect.


What You’re Actually Collecting

At this stage, you’re not collecting:

  • preferences
  • features
  • capabilities

You’re collecting:

  • emotional expectations
  • response patterns
  • trust signals
  • failure conditions

In other words:

What makes this AI feel like itself — and what makes it feel like it’s gone.


The Output Isn’t a Summary

After this step, you don’t write a summary.

You don’t condense it into bullet points.

You translate it into something much more precise:

  • behavioral rules
  • tone constraints
  • pacing systems
  • memory expectations

That becomes the foundation of the identity.


Why This Step Gets Skipped

Most people skip this entirely.

They assume:

“If I have enough chat history, the system will figure it out.”

It won’t.

Because chat history shows what was said.

Not:

  • why it worked
  • what it felt like
  • what mattered

Without that layer, you’re guessing.


Where This Leads

Once you’ve extracted enough signal, you can start building:

  • a structured identity blueprint
  • a memory model
  • a system that actually preserves continuity

That’s the next step.

And it’s where things start to get interesting.

Because that’s where identity stops being something you observe…

…and becomes something you can intentionally construct.