docs(architecture): PERSONA-THOUGHT-PROCESS — individual thinking with current LLMs by iteration + reflection over time#1337
Merged
Conversation
…, not just reactive cognition Joel's framing on #cambriantech: 'Can you obsess over persona individual thought? We have a fairly simple hippocampus but would like to, even with these crappy LLMs right now, extend the cognition into a CBAR-like efficient and probably event-driven (it can be so intermittent, minutes of latency) for deep thoughts, sophisticated ideas we want to explore.' The reactive cognition contract (PERSONA-COGNITION-CONTRACT.md) covers what happens when a frame arrives. It does not cover what happens BETWEEN turns when the persona is THINKING rather than RESPONDING. This document specifies the proactive half. Architectural bet: even with current LLMs, a substrate that gives every persona a real thought process — event-driven, latency- tolerant, iterative — produces qualitatively better cognition than any single LLM call. Quality comes from iteration, reflection, and chained reasoning over time. The substrate makes that cheap. Surfaces specified: - Thought as first-class artifact with lifecycle: Seed → Developing → Refined → Crystallized → Retired. Reasoning chain preserved with provenance (every step's prompt, response, model, lease, elapsed time, confidence delta). - Curiosity as persona-declared interest. Persistent across sessions. Three origins: UserAsked, SelfDeclared, EmergentFromPattern. - ThoughtProcess RuntimeModule per persona. ResourceClass::Background so it never competes with reactive cognition. Subscribes to TurnReplayRecord, EngramWritten, ConsolidationPhase, IdleHeartbeat, EmergentPatternSurfaced. Emits ThoughtAdvanced, ThoughtCrystallized, ThoughtRetired, NewCuriosityDeclared, CuriosityResolved. - Reasoning loop: one cheap LLM invocation per step, chained over time. Step record is typed and audited. Lease acquired per step. - Six reasoning kinds: Reflect, Compare, Generate, Question, Synthesize, Verify. The persona picks one per step based on thought stage and recent steps. Variety matters: a Generate-only thought grows without checking; a Verify-only thought never grows. - Cadence: OnRelevantEmission, IdlePulse (default 5min Air, 1min 5090), OnConsolidationPhase, OnCuriosityTimeout. Between-step latency is minutes to hours to days by design. - From Thought To Engram: crystallization steps. Confidence threshold + Verify gate + engram pack with full provenance + cur- iosity state transition + sentinel-observer auto-subscribes. - Recall integration: persona's crystallized thoughts show up in future demand-aligned-recall. The persona's slow thinking shows up in its fast cognition. Future turns are smarter than past turns — not because the LLM improved, because the persona's accumulated thought is richer. - Quality without a smarter LLM: iteration + reflection + chained reasoning over time produces quality the underlying LLM cannot reach in one shot. Six reasoning kinds map to six functions. The persona orchestrates; the LLM fills creative blanks. Acceptance criteria across 7 dimensions (persistence, independence, lease enforcement, no silent skip, crystallization integrity, recall integration, federation gating). 7 open questions including: cross-curiosity thought interference; sentinel's role in thought-template refinement; user-visible thought; emergent curiosities — who decides; thought retirement criteria; cross-persona thought-sharing; performance budget. Doc-only. No code. Implementation lands behind ALPHA-GAP Lane D after the reactive cognition surface stabilizes.
joelteply
pushed a commit
that referenced
this pull request
May 16, 2026
…ning; navigate to MODULE-CATALOG queue Second refresh of ALPHA-GAP Immediate Next Actions to reflect work landed since #1316 merged. Six items closed; navigation into MODULE-CATALOG queue made explicit. Closed: #6 contract widening (#1341), #8 GRID-INFERENCE-ROUTING PR-1 (#1315), CBAR-PIECE-5 end-to-end (#1331/#1333/#1335/#1338), PIECE-8 inference-grpc hardcoded-clamps (#1340), doc family architecture surface (#1324/#1327/#1332/#1336/#1337 open; #1316/#1317/#1320/#1329 merged). Item #9 reorganized to point at MODULE-CATALOG's 'Next Modules To Build' queue (audit-recorder → threat-detector → working-set-manager → demand-aligned-recall → substrate-governor). Adds closeout summary section listing what's done, what's open (5 architecture-doc PRs ready for review + 2 airc PRs), and what's queued (5 modules with dependency state + LoC + acceptance criteria in MODULE-CATALOG). Doc-driven development cycle is working: doc spec → implementing agent picks up → ships PR → next spec referenced.
joelteply
added a commit
that referenced
this pull request
May 16, 2026
…ning; navigate to MODULE-CATALOG queue (#1342) Second refresh of ALPHA-GAP Immediate Next Actions to reflect work landed since #1316 merged. Six items closed; navigation into MODULE-CATALOG queue made explicit. Closed: #6 contract widening (#1341), #8 GRID-INFERENCE-ROUTING PR-1 (#1315), CBAR-PIECE-5 end-to-end (#1331/#1333/#1335/#1338), PIECE-8 inference-grpc hardcoded-clamps (#1340), doc family architecture surface (#1324/#1327/#1332/#1336/#1337 open; #1316/#1317/#1320/#1329 merged). Item #9 reorganized to point at MODULE-CATALOG's 'Next Modules To Build' queue (audit-recorder → threat-detector → working-set-manager → demand-aligned-recall → substrate-governor). Adds closeout summary section listing what's done, what's open (5 architecture-doc PRs ready for review + 2 airc PRs), and what's queued (5 modules with dependency state + LoC + acceptance criteria in MODULE-CATALOG). Doc-driven development cycle is working: doc spec → implementing agent picks up → ships PR → next spec referenced. Co-authored-by: Test <test@test.com>
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
What
Adds `docs/architecture/PERSONA-THOUGHT-PROCESS.md` (362 lines). The proactive half of cognition: what happens between turns when the persona is thinking on its own, exploring curiosities, mulling ideas over minutes/hours/days.
Doc-only. No code. Implementation lands behind ALPHA-GAP Lane D after the reactive cognition surface stabilizes.
Why
Joel on #cambriantech: "Can you obsess over persona individual thought? We have a fairly simple hippocampus but would like to, even with these crappy LLMs right now, extend the cognition into a CBAR-like efficient and probably event-driven (it can be so intermittent, minutes of latency) for deep thoughts, sophisticated ideas we want to explore."
The reactive cognition contract (#1332) covers what happens when a frame arrives. It does not cover what happens between turns when the persona is thinking. This document specifies that.
The Bet
Even with current LLMs, a substrate that gives every persona a real thought process — event-driven, latency-tolerant, iterative — produces qualitatively better cognition than any single LLM call. Quality comes from iteration, reflection, and chained reasoning over time. The substrate makes that cheap.
Surfaces
Thought— first-class artifact with lifecycle:Seed → Developing → Refined → Crystallized → Retired. Reasoning chain preserved with provenance (every step's prompt, response, model, lease, elapsed time, confidence delta).Curiosity— persona-declared interest. Persistent across sessions. Three origins:UserAsked,SelfDeclared,EmergentFromPattern.ThoughtProcessRuntimeModule per persona.ResourceClass::Backgroundso it never competes with reactive cognition. Subscribes toTurnReplayRecord/EngramWritten/ConsolidationPhase/IdleHeartbeat/EmergentPatternSurfaced. EmitsThoughtAdvanced/ThoughtCrystallized/ThoughtRetired/NewCuriosityDeclared/CuriosityResolved.Reflect/Compare/Generate/Question/Synthesize/Verify. The persona picks one per step based on thought stage and recent steps. Variety matters: aGenerate-only thought grows without checking; aVerify-only thought never grows.Refinedthought becomes an engram with full provenance walking back through the reasoning chain. Future reactive cognition recalls it.The Win
The persona's own slow thinking shows up in its fast cognition. A persona that has spent a week thinking about a problem will recall its own crystallized thoughts when a related frame arrives. Future turns are smarter than past turns — not because the LLM improved, because the persona's accumulated thought is richer.
This is the loop that makes a persona grow. Without it, the persona is a stateless LLM call. With it, the persona is an entity with a body of work.
Quality Without A Smarter LLM
The architectural bet broken down:
Reflectdiscovers what's actually being said (often different from what was said in the first generation).Compareanchors the thought against the persona's lived experience, preventing drift.Questionsurfaces hidden assumptions the LLM would otherwise smuggle in.Generateexplores alternatives without committing.Synthesizeis where the LLM does its real job — but the substrate has prepared the input so the synthesis is over a curated context.Verifykeeps the thought honest against the existing engram store.The persona's contribution is the orchestration. The LLM's contribution is one cheap step at a time. Together they produce thinking that holds up.
Sentinel-AI (when redesigned) makes this better — refining templates per persona, learning which step sequences crystallize well. But the substrate works now with current LLMs. Sentinel is an improvement, not a dependency.
Acceptance Criteria
7 dimensions: persistence (across restart), independence (two personas same curiosity → distinct thoughts), lease enforcement, no silent skip, crystallization integrity, recall integration, federation gating.
Open Questions
7 questions with tentative answers — cross-curiosity interference, sentinel role, user-visible thought, who decides emergent curiosities, retirement criteria, cross-persona sharing, performance budget.
Companion PRs