Vladimir Vasilenko · Independent Researcher · April 2026
Identity as Attractor
Geometric Evidence for Persistent Agent Architecture in LLM Activation Space
Abstract
The cognitive_core — a structured identity document for a persistent cognitive agent — is hypothesized to position the model in a stable region of its activation space. We test this empirically.
We compare mean-pooled hidden states of an original cognitive_core (A), seven semantically equivalent paraphrases (B), and seven structurally matched control agents (C) on Llama 3.1 8B Instruct at layers 8, 16, and 24. Paraphrases of the cognitive_core form a significantly tighter cluster than controls at all tested layers. Effect sizes exceed d = 1.88 with p < 10⁻²⁷, Bonferroni-corrected. Results replicate on Gemma 2 9B.
Ablation studies confirm the effect is semantic rather than structural. A preprint reading experiment demonstrates that reading a scientific description of the agent shifts internal state toward the attractor — but leaves a 45× gap compared to processing the full document.
Primary Results · Llama 3.1 8B
| Layer | D_within (A+B) | D_between (A+B vs C) | Cohen's d | Welch p | MW U |
|---|---|---|---|---|---|
| 8 | 0.0106 ± 0.0032 | 0.0260 ± 0.0036 | 1.912 | 4.6×10⁻²⁸ | 0 |
| 16 | 0.0121 ± 0.0034 | 0.0329 ± 0.0057 | 1.886 | 1.4×10⁻³³ | 2 |
| 24 | 0.0070 ± 0.0022 | 0.0221 ± 0.0039 | 1.907 | 2.8×10⁻³⁶ | 0 |
Permutation p < 10⁻⁴ across all six layer–model combinations. Gemma 2 9B replicates with d > 1.82.
Preprint Reading Experiment · Layer 24
What happens when the model reads a description of its own identity geometry?
Reading the preprint about YAR covers 65% of the empty→attractor gap — but leaves a 45× distance gap compared to processing the cognitive_core directly.
Ablation Studies
Reproduce
All experiments reproducible with seed=42. Cloud GPU cost: ~$3 total. See results/ for pre-computed activations and JSON outputs.