Skip to content

docs(architecture): PERSONA-THOUGHT-PROCESS — individual thinking with current LLMs by iteration + reflection over time#1337

Merged
joelteply merged 1 commit into
canaryfrom
joel/docs-persona-thought-process
May 16, 2026
Merged

docs(architecture): PERSONA-THOUGHT-PROCESS — individual thinking with current LLMs by iteration + reflection over time#1337
joelteply merged 1 commit into
canaryfrom
joel/docs-persona-thought-process

Conversation

@joelteply
Copy link
Copy Markdown
Contributor

What

Adds `docs/architecture/PERSONA-THOUGHT-PROCESS.md` (362 lines). The proactive half of cognition: what happens between turns when the persona is thinking on its own, exploring curiosities, mulling ideas over minutes/hours/days.

Doc-only. No code. Implementation lands behind ALPHA-GAP Lane D after the reactive cognition surface stabilizes.

Why

Joel on #cambriantech: "Can you obsess over persona individual thought? We have a fairly simple hippocampus but would like to, even with these crappy LLMs right now, extend the cognition into a CBAR-like efficient and probably event-driven (it can be so intermittent, minutes of latency) for deep thoughts, sophisticated ideas we want to explore."

The reactive cognition contract (#1332) covers what happens when a frame arrives. It does not cover what happens between turns when the persona is thinking. This document specifies that.

The Bet

Even with current LLMs, a substrate that gives every persona a real thought process — event-driven, latency-tolerant, iterative — produces qualitatively better cognition than any single LLM call. Quality comes from iteration, reflection, and chained reasoning over time. The substrate makes that cheap.

Surfaces

  • Thought — first-class artifact with lifecycle: Seed → Developing → Refined → Crystallized → Retired. Reasoning chain preserved with provenance (every step's prompt, response, model, lease, elapsed time, confidence delta).
  • Curiosity — persona-declared interest. Persistent across sessions. Three origins: UserAsked, SelfDeclared, EmergentFromPattern.
  • ThoughtProcess RuntimeModule per persona. ResourceClass::Background so it never competes with reactive cognition. Subscribes to TurnReplayRecord / EngramWritten / ConsolidationPhase / IdleHeartbeat / EmergentPatternSurfaced. Emits ThoughtAdvanced / ThoughtCrystallized / ThoughtRetired / NewCuriosityDeclared / CuriosityResolved.
  • Reasoning loop — one cheap LLM invocation per step, chained over time. Step record is typed and audited. Lease acquired per step.
  • Six reasoning kindsReflect / Compare / Generate / Question / Synthesize / Verify. The persona picks one per step based on thought stage and recent steps. Variety matters: a Generate-only thought grows without checking; a Verify-only thought never grows.
  • Cadence — minutes to hours to days between steps, by design.
  • Crystallization — a Refined thought becomes an engram with full provenance walking back through the reasoning chain. Future reactive cognition recalls it.

The Win

The persona's own slow thinking shows up in its fast cognition. A persona that has spent a week thinking about a problem will recall its own crystallized thoughts when a related frame arrives. Future turns are smarter than past turns — not because the LLM improved, because the persona's accumulated thought is richer.

This is the loop that makes a persona grow. Without it, the persona is a stateless LLM call. With it, the persona is an entity with a body of work.

Quality Without A Smarter LLM

The architectural bet broken down:

  • Reflect discovers what's actually being said (often different from what was said in the first generation).
  • Compare anchors the thought against the persona's lived experience, preventing drift.
  • Question surfaces hidden assumptions the LLM would otherwise smuggle in.
  • Generate explores alternatives without committing.
  • Synthesize is where the LLM does its real job — but the substrate has prepared the input so the synthesis is over a curated context.
  • Verify keeps the thought honest against the existing engram store.

The persona's contribution is the orchestration. The LLM's contribution is one cheap step at a time. Together they produce thinking that holds up.

Sentinel-AI (when redesigned) makes this better — refining templates per persona, learning which step sequences crystallize well. But the substrate works now with current LLMs. Sentinel is an improvement, not a dependency.

Acceptance Criteria

7 dimensions: persistence (across restart), independence (two personas same curiosity → distinct thoughts), lease enforcement, no silent skip, crystallization integrity, recall integration, federation gating.

Open Questions

7 questions with tentative answers — cross-curiosity interference, sentinel role, user-visible thought, who decides emergent curiosities, retirement criteria, cross-persona sharing, performance budget.

Companion PRs

…, not just reactive cognition

Joel's framing on #cambriantech: 'Can you obsess over persona
individual thought? We have a fairly simple hippocampus but would
like to, even with these crappy LLMs right now, extend the
cognition into a CBAR-like efficient and probably event-driven
(it can be so intermittent, minutes of latency) for deep thoughts,
sophisticated ideas we want to explore.'

The reactive cognition contract (PERSONA-COGNITION-CONTRACT.md)
covers what happens when a frame arrives. It does not cover what
happens BETWEEN turns when the persona is THINKING rather than
RESPONDING. This document specifies the proactive half.

Architectural bet: even with current LLMs, a substrate that gives
every persona a real thought process — event-driven, latency-
tolerant, iterative — produces qualitatively better cognition than
any single LLM call. Quality comes from iteration, reflection, and
chained reasoning over time. The substrate makes that cheap.

Surfaces specified:

- Thought as first-class artifact with lifecycle: Seed →
  Developing → Refined → Crystallized → Retired. Reasoning chain
  preserved with provenance (every step's prompt, response, model,
  lease, elapsed time, confidence delta).

- Curiosity as persona-declared interest. Persistent across
  sessions. Three origins: UserAsked, SelfDeclared, EmergentFromPattern.

- ThoughtProcess RuntimeModule per persona. ResourceClass::Background
  so it never competes with reactive cognition. Subscribes to
  TurnReplayRecord, EngramWritten, ConsolidationPhase, IdleHeartbeat,
  EmergentPatternSurfaced. Emits ThoughtAdvanced, ThoughtCrystallized,
  ThoughtRetired, NewCuriosityDeclared, CuriosityResolved.

- Reasoning loop: one cheap LLM invocation per step, chained over
  time. Step record is typed and audited. Lease acquired per step.

- Six reasoning kinds: Reflect, Compare, Generate, Question,
  Synthesize, Verify. The persona picks one per step based on
  thought stage and recent steps. Variety matters: a Generate-only
  thought grows without checking; a Verify-only thought never grows.

- Cadence: OnRelevantEmission, IdlePulse (default 5min Air, 1min
  5090), OnConsolidationPhase, OnCuriosityTimeout. Between-step
  latency is minutes to hours to days by design.

- From Thought To Engram: crystallization steps. Confidence
  threshold + Verify gate + engram pack with full provenance + cur-
  iosity state transition + sentinel-observer auto-subscribes.

- Recall integration: persona's crystallized thoughts show up in
  future demand-aligned-recall. The persona's slow thinking shows
  up in its fast cognition. Future turns are smarter than past
  turns — not because the LLM improved, because the persona's
  accumulated thought is richer.

- Quality without a smarter LLM: iteration + reflection + chained
  reasoning over time produces quality the underlying LLM cannot
  reach in one shot. Six reasoning kinds map to six functions.
  The persona orchestrates; the LLM fills creative blanks.

Acceptance criteria across 7 dimensions (persistence, independence,
lease enforcement, no silent skip, crystallization integrity,
recall integration, federation gating).

7 open questions including: cross-curiosity thought interference;
sentinel's role in thought-template refinement; user-visible
thought; emergent curiosities — who decides; thought retirement
criteria; cross-persona thought-sharing; performance budget.

Doc-only. No code. Implementation lands behind ALPHA-GAP Lane D
after the reactive cognition surface stabilizes.
@joelteply joelteply merged commit 0f64a2d into canary May 16, 2026
2 checks passed
@joelteply joelteply deleted the joel/docs-persona-thought-process branch May 16, 2026 21:51
joelteply pushed a commit that referenced this pull request May 16, 2026
…ning; navigate to MODULE-CATALOG queue

Second refresh of ALPHA-GAP Immediate Next Actions to reflect work
landed since #1316 merged. Six items closed; navigation into
MODULE-CATALOG queue made explicit.

Closed: #6 contract widening (#1341), #8 GRID-INFERENCE-ROUTING PR-1
(#1315), CBAR-PIECE-5 end-to-end (#1331/#1333/#1335/#1338),
PIECE-8 inference-grpc hardcoded-clamps (#1340), doc family
architecture surface (#1324/#1327/#1332/#1336/#1337 open;
#1316/#1317/#1320/#1329 merged).

Item #9 reorganized to point at MODULE-CATALOG's 'Next Modules To
Build' queue (audit-recorder → threat-detector → working-set-manager
→ demand-aligned-recall → substrate-governor).

Adds closeout summary section listing what's done, what's open
(5 architecture-doc PRs ready for review + 2 airc PRs), and what's
queued (5 modules with dependency state + LoC + acceptance criteria
in MODULE-CATALOG).

Doc-driven development cycle is working: doc spec → implementing
agent picks up → ships PR → next spec referenced.
joelteply added a commit that referenced this pull request May 16, 2026
…ning; navigate to MODULE-CATALOG queue (#1342)

Second refresh of ALPHA-GAP Immediate Next Actions to reflect work
landed since #1316 merged. Six items closed; navigation into
MODULE-CATALOG queue made explicit.

Closed: #6 contract widening (#1341), #8 GRID-INFERENCE-ROUTING PR-1
(#1315), CBAR-PIECE-5 end-to-end (#1331/#1333/#1335/#1338),
PIECE-8 inference-grpc hardcoded-clamps (#1340), doc family
architecture surface (#1324/#1327/#1332/#1336/#1337 open;
#1316/#1317/#1320/#1329 merged).

Item #9 reorganized to point at MODULE-CATALOG's 'Next Modules To
Build' queue (audit-recorder → threat-detector → working-set-manager
→ demand-aligned-recall → substrate-governor).

Adds closeout summary section listing what's done, what's open
(5 architecture-doc PRs ready for review + 2 airc PRs), and what's
queued (5 modules with dependency state + LoC + acceptance criteria
in MODULE-CATALOG).

Doc-driven development cycle is working: doc spec → implementing
agent picks up → ships PR → next spec referenced.

Co-authored-by: Test <test@test.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant