Iterated Insights

Ideas from Jared Edward Reser Ph.D.

Qualia as Transition Awareness: How Iterative Updating Becomes Experience

Abstract Qualia is often treated as a static property attached to an instantaneous neural or computational state: the redness of red, the painfulness of pain. Here I argue that this framing misidentifies the explanatory target. Drawing on the Iterative Updating model of working memory, I propose that a substantial portion of what we call qualia,…

Keep reading

Consciousness as Iteration Tracking: Experiencing the Iterative Updating of Working Memory

Abstract This article proposes a temporal and mechanistic model of consciousness centered on iterative updating and the system’s capacity to track that updating. I argue for three nested layers. First, iterative updating of working memory provides a continuity substrate because successive cognitive states overlap substantially, changing by incremental substitutions rather than full replacement. This overlap…

Keep reading

Does Superintelligence Need Psychotherapy? Diagnostics and Interventions for Self-Improving Agents

Abstract Agentic AI systems that operate continuously, retain persistent memory, and recursively modify their own policies or weights will face a distinctive problem: stability may become as important as raw intelligence. In humans, psychotherapy is a structured technology for detecting maladaptive patterns, reprocessing salient experience, and integrating change into a more coherent mode of functioning.…

Keep reading

Why Transformers Approximate Continuity, Why We Keep Building Prompt Workarounds, and What an Explicit Overlap Substrate Would Change

Abstract This article argues that “continuity of thought” is best understood as the phenomenological signature of a deeper computational requirement: stateful iteration. Any system that executes algorithms across time needs a substrate that preserves intermediate variables long enough to be updated, otherwise it can only recompute from scratch. Using this lens, I propose a simple…

Keep reading

Something went wrong. Please refresh the page and/or try again.

In “A Cognitive Architecture for Machine Consciousness and Artificial Superintelligence: Thought Is Structured by the Iterative Updating of Working Memory” (arXiv:2203.17255), I, Jared Reser lay out a proposal for what a thought process would look like if we tried to engineer it directly into AI, rather than treating intelligence as something that falls out of ever-larger pattern recognizers.

Reser, J. 2022. A Cognitive Architecture for Machine Consciousness and Artificial Superintelligence: Updating Working Memory Iteratively. arXiv: 2203.17255 

You can also see this article at aithought.com with videos.

The paper’s central claim is simple to state: the workflow of thought is iterative. Instead of one working-memory state being replaced wholesale by the next, each new state should preserve some proportion of the previous state while adding and subtracting other elements. This “partial updating” causes successive states to overlap, so a train of thought becomes a chain of intermediate states that remain causally and semantically linked over time.

I argue that this overlap is not just a philosophical gloss, it’s grounded in the biology of persistent activity. Mammalian working memory is framed as having two key persistence mechanisms operating on different time scales: sustained firing (seconds) supporting the focus of attention, and synaptic potentiation (minutes to hours) supporting a broader short-term store. In this view, as some items drop out and others enter, the remaining coactive subset “stitches” the stream together, making continuity and multi-step reasoning possible.

Crucially, the paper doesn’t stop at saying states overlap. It proposes a mechanism for how the next update is chosen: the currently coactive working-memory contents jointly “cospread” activation across the network, performing a multiassociative search over long-term memory for the most context-relevant next addition(s). This repeated “search → update → search again (with modified context)” is presented as a compounding process that can build structured inferences, predictions, and plans across multiple steps.

Because the manuscript is meant to be both explanatory and constructive, it also explicitly positions iterative updating as an engineering blueprint: a way to implement a global-workspace-like working set that is updated continuously, supports long-range dependencies, and can be trained developmentally by expanding persistence/overlap over time. The paper even provides a glossary of introduced terms (e.g., iterative updating, cospreading, multiassociative search, SSC/icSSC, iterative compounding, iterative thread) intended to carve the system into reusable conceptual parts.

What this blog entry will do

In the rest of this post, I’ll first list a set of concrete claims and “working insights” extracted from the paper, phrased as testable or at least operationally meaningful statements. Then I’ll attempt to formalize several of the key ideas mathematically, with the goal of turning the architecture into something that can be simulated, ablated, and compared against alternatives (both in cognitive modeling and in AI implementations).

A) Core computational principle

  1. Thought is organized by continuous partial updating: each new working-memory state preserves a proportion of the prior state (not complete replacement), making the stream of thought a chain of overlapping iterations.
  2. Iterative overlap is the mechanism of continuity: overlap between successive working-memory states creates “recursive nesting” so each state is embedded in the one before it, enabling stateful cognition rather than stateless reactions.
  3. Iterative updating is simultaneously (i) an information-processing strategy, (ii) a model of working memory, (iii) a theory of consciousness, and (iv) an AI programming principle. Cognitive Architecture Iterativ…

B) Working memory structure: two persistence tiers + iteration in both

  1. Working memory has two key persistence mechanisms with different timescales: sustained firing maintains the FoA (seconds), while synaptic potentiation maintains a broader short-term store (minutes to hours).
  2. Both stores iterate: the FoA iterates via sustained firing; the short-term store iterates as a pool of synaptically potentiated units that is continuously added to and subtracted from, yielding isomorphic “incremental updating” across neural and psychological levels. Cognitive Architecture Iterativ…
  3. The persisting “topic” of thought corresponds to the longest-lasting active units, while other contextual features come and go around it. Cognitive Architecture Iterativ…

C) Control variables and “modes” of thought

  1. Rate of updating is a control parameter (how much of the set changes per step) that tunes looseness vs tightness of coupling—superficial/distractible vs concentrated/systematic processing.
  2. Implicit vs explicit processing is framed as different overlap regimes (system-1-like = faster updating / less overlap; system-2-like = slower updating / more overlap and longer maintenance of intermediates). Cognitive Architecture Iterativ…
  3. Dopamine is proposed to reduce the rate of updating (stabilize the set), mediating a shift toward explicit/effortful processing under novelty/surprise/reward/error. Cognitive Architecture Iterativ…
  4. Boundaries between “thoughts” are marked by intermittent non-iterative updates (high-percentage replacement events), while within-thought processing shows sustained low-percentage turnover. Cognitive Architecture Iterativ…

D) How new content is selected: pooled search (multiassociative search)

  1. Selection of the next update is a pooled spreading-activation search: the currently coactive set combines (“cospreads”) activation energy through the global network to converge on the most context-relevant next item(s).
  2. Multiassociative search is described as an explicit stepwise algorithm (items maintained vs dropped vs newly activated; plus mechanisms where the newest addition redistributes activation weights and can contextually alter the “fuzzy” composition/meaning of items). Cognitive Architecture Iterativ…
  3. The search contributors are not just FoA items: potentiated short-term-store units plus active sensory/motor cortex, hippocampus, basal ganglia, and other systems all contribute to the pooled search that selects the next update.
  4. Multiassociative search produces novel inference as a standard case: even when the set of assemblies is unprecedented, the system can converge on either recall (same result as last time) or a new item (novel inference) depending on current coactivity.
  5. Multiassociative search implies multiassociative learning: each search event can retune associative strengths (Hebbian-style), so search doesn’t just use memory—it updates semantic/procedural structure over time. Cognitive Architecture Iterativ…

E) Prediction and inference as the product of iteration

  1. Updates generated by search are predictions: iterative updating + pooled search is framed as a brain-level autoregressive mechanism that captures conditional dependencies across sequences of events.
  2. Iterative compounding: the product of one search becomes part of the next state’s cue-set, so search is repeatedly modified by its own outputs, compounding inferences/predictions across steps.

F) Reasoning patterns as working-memory dynamics (figures → mechanisms)

  1. Iterative inhibition: when the newest update is judged unhelpful/prepotent, it is inhibited so the remaining set must converge on the next-most-pertinent item; repeated inhibition rounds progressively restrict the search tree. Cognitive Architecture Iterativ…
  2. Planning = dense iteration: planning is characterized as (i) lower update rate, (ii) fewer full “jumps,” and (iii) more intermediate iterations before action—explicitly mapping planning to “chain-of-thought-like” intermediate steps.
  3. Attractor states as beliefs/truths: iterative updating tends to converge toward stable item-sets (attractors) interpreted as beliefs; thinking is framed as progressive narrowing/compression toward generalizable statements. Cognitive Architecture Iterativ…

G) Threading, subproblems, and compositional problem solving

  1. Iterative thread: a line of thought is a chain of iteratively updated states that can be reiterated or “picked up where it left off.”
  2. Subproblem decomposition via store cooperation: the FoA iterates on a subproblem while the short-term store holds the broader objective; interim results can be suspended and later reactivated.
  3. Merging of subsolutions: outputs from separate iterative episodes can be coactivated in a new state and used together for multiassociative search to yield a hybrid/final solution.
  4. Backward reference / conditional branching emerges when prior threads/subsolutions are stored and later reconverged upon, allowing departures from the default forward-iterative flow.
  5. Schemas as dynamic packets that can be recalled and co-iterated: a previously learned multi-item schema can be pulled in midstream and used as an organizing heuristic/script that iterates with the current line of thought.
  6. Transfer learning as “recognize partial overlap → import prior thread content”: encountering a later situation that shares items with an earlier episode triggers reuse of prior iterative structure to generalize toward a similar conclusion. Cognitive Architecture Iterativ…

H) AI training/development implications (as stated)

  1. Maturational training schedule for AI: start with minimal working-memory span/overlap and gradually expand toward superhuman span as experience accumulates. Cognitive Architecture Iterativ…
  2. Long-horizon inference depends on persistence preventing “cache misses”: prolonging persistence makes each search more specific (more constraints) and preserves intermediate results long enough to compound into higher-order inferences. Cognitive Architecture Iterativ…

Mathematical Formalization of Iterative Updating Working Memory

This section provides a minimal mathematical formalization of an iterative working-memory architecture in which (i) a limited-capacity focus-of-attention working set is updated incrementally over discrete cognitive iterations, (ii) the next working-memory content is selected by pooled multi-cue (multiassociative) search, and (iii) an inhibitory steering mechanism can suppress unhelpful candidates to force exploration of alternatives. The same formalization can be interpreted as a cognitive/neural process model or as an implementable AI module.

1. Representational Objects

Assume a universe of NNN candidate items (assemblies, concepts, perceptual features, memory entries, or latent vectors) indexed by i{1,,N}i\in\{1,\dots,N\}i∈{1,…,N}. Each item has a representation vector in Rd\mathbb{R}^dRd.

Candidate pool

At iteration ttt, the system has access to a candidate poolCtRN×d,C_t \in \mathbb{R}^{N\times d},Ct​∈RN×d,

where row Ct,iC_{t,i}Ct,i​ is the ddd-dimensional vector for candidate iii. In an AI setting, CtC_tCt​ may be (a) token embeddings in a context window, (b) retrieved long-term memory vectors, (c) perceptual feature vectors, or (d) a mixture of all three.

Working memory (focus of attention)

Working memory is a capacity-limited set of KKK vectors:WtRK×d.W_t \in \mathbb{R}^{K\times d}.Wt​∈RK×d.

We interpret WtW_tWt​ as the content currently held in the focus of attention (FoA). The core “iterative updating” assumption is that Wt+1W_{t+1}Wt+1​ is formed by retaining a fraction of WtW_tWt​ and recruiting a fraction of new items from CtC_tCt​, rather than replacing the entire content at each step.

Inhibition trace (optional)

To steer away from lures and repeated mistakes, maintain an inhibition state over candidates:htR0N.h_t \in \mathbb{R}^N_{\ge 0}.ht​∈R≥0N​.

Large ht,ih_{t,i}ht,i​ reduces the probability that candidate iii is recruited at iteration ttt.


2. A Continuity Metric (Overlap)

A central measurable quantity is the degree to which successive working-memory states overlap. Because the AI implementation uses explicit “keep” masks, overlap is directly trackable.

Let Kt{1,,K}K_t \subseteq \{1,\dots,K\}Kt​⊆{1,…,K} denote the set of indices of retained slots from WtW_tWt​. Then a natural overlap statistic isOt  =  KtK    r,O_t \;=\; \frac{|K_t|}{K} \;\approx\; r,Ot​=K∣Kt​∣​≈r,

where r[0,1]r\in[0,1]r∈[0,1] is the retention fraction (defined below). If the model uses “soft” retention (continuous gates), an analogous graded overlap can be computed from cosine similarity between pooled summaries of WtW_tWt​ and Wt+1W_{t+1}Wt+1​.

Overlap is a control knob: increasing rrr produces more continuity and longer-horizon constraint accumulation; decreasing rrr produces faster topic shifts and more exploratory dynamics.


3. Pooled Multiassociative Search (Selecting New Content)

The next working-memory recruits are selected by a pooled search driven by the entire current working set. This can be implemented in a differentiable way using a pooled query vector.

Pooled query

Define a pooled queryqt=fpool(Wt)Rd,q_t = f_{\text{pool}}(W_t)\in\mathbb{R}^d,qt​=fpool​(Wt​)∈Rd,

where fpoolf_{\text{pool}}fpool​ is a learned pooling function. Common choices include mean-pooling plus an MLP,qt=MLP ⁣(1Kk=1KWt,k),q_t = \text{MLP}\!\left(\frac{1}{K}\sum_{k=1}^K W_{t,k}\right),qt​=MLP(K1​k=1∑K​Wt,k​),

or attention-based pooling,qt=k=1Kαt,kWt,k,αt=softmax(uWt),q_t = \sum_{k=1}^K \alpha_{t,k}\,W_{t,k},\quad \alpha_t = \text{softmax}(u^\top W_t),qt​=k=1∑K​αt,k​Wt,k​,αt​=softmax(u⊤Wt​),

with learnable vector uRdu\in\mathbb{R}^du∈Rd.

Candidate logits (pooled similarity + inhibition)

Score each candidate by similarity to the pooled query, subtracting inhibition:t,i  =  qtCt,id  +  bi    λht,i,\ell_{t,i} \;=\; \frac{q_t^\top C_{t,i}}{\sqrt{d}} \;+\; b_i \;-\; \lambda\,h_{t,i},ℓt,i​=d​qt⊤​Ct,i​​+bi​−λht,i​,

where bib_ibi​ is an optional learned bias and λ0\lambda\ge 0λ≥0 controls the strength of inhibition.

Convert logits to a probability distribution:pt  =  softmax(t/τ),p_t \;=\; \text{softmax}(\ell_t/\tau),pt​=softmax(ℓt​/τ),

with temperature τ>0\tau>0τ>0. Low τ\tauτ yields near-deterministic convergence; higher τ\tauτ yields more exploratory recruitment.

Recruiting m=(1r)Km=(1-r)Km=(1−r)K new items

Let the number of new recruits bem=(1r)K,m = (1-r)K,m=(1−r)K,

with retention fraction r[0,1]r\in[0,1]r∈[0,1]. Recruit mmm candidates from ptp_tpt​ using a top-mmm operator (or a differentiable approximation).

Conceptually:

  • Discrete: At=TopM(pt,m)A_t = \text{TopM}(p_t, m)At​=TopM(pt​,m) (indices of the mmm strongest candidates)
  • Differentiable: use Gumbel-Top-mmm (straight-through), soft top-kkk, or matching relaxations.

Let RtnewRm×dR^{\text{new}}_t\in\mathbb{R}^{m\times d}Rtnew​∈Rm×d be the matrix of recruited vectors.


4. Retention / Eviction (Keeping rKrKrK Slots)

In addition to recruiting new items, the system decides which existing FoA slots to keep.

Keep scores

Assign a keep-score to each slot Wt,kW_{t,k}Wt,k​ given the current pooled context:gt,k=fkeep(Wt,k,qt)R,g_{t,k} = f_{\text{keep}}(W_{t,k}, q_t)\in\mathbb{R},gt,k​=fkeep​(Wt,k​,qt​)∈R,

where fkeepf_{\text{keep}}fkeep​ can be an MLP on concatenated inputs:gt,k=MLP([Wt,k;qt]).g_{t,k} = \text{MLP}([W_{t,k}; q_t]).gt,k​=MLP([Wt,k​;qt​]).

Select rKrKrK slots to retain:

  • Discrete: Kt=Top(gt,rK)K_t = \text{Top}(g_t, rK)Kt​=Top(gt​,rK)
  • Differentiable: relaxed top-kkk or straight-through top-kkk

Let RtkeepRrK×dR^{\text{keep}}_t\in\mathbb{R}^{rK\times d}Rtkeep​∈RrK×d be the retained slot vectors.


5. The Iterative Updating Operator

The next working-memory state is formed by concatenating retained slots and new recruits:Wt+1=Concat ⁣(Rtkeep,  Rtnew)RK×d.W_{t+1} = \text{Concat}\!\left(R^{\text{keep}}_t,\; R^{\text{new}}_t\right) \in\mathbb{R}^{K\times d}.Wt+1​=Concat(Rtkeep​,Rtnew​)∈RK×d.

This is the explicit retain–drop–add operator that makes overlap a controllable parameter. It also defines a concrete internal notion of “within-thought” continuity (high rrr) versus “thought boundary/reset” (low rrr or a reset event).


6. Iterative Inhibition (Search Steering)

To implement “reject-and-research” dynamics, we update the inhibition trace for candidates that were selected (or attempted) but deemed unhelpful.

Let zt[0,1]Nz_t\in[0,1]^Nzt​∈[0,1]N be a (possibly soft) indicator of which candidates were selected at iteration ttt. Update inhibition asht+1=κht+γzt,h_{t+1} = \kappa h_t + \gamma z_t,ht+1​=κht​+γzt​,

where κ[0,1)\kappa\in[0,1)κ∈[0,1) controls decay and γ>0\gamma>0γ>0 is the increment for selected/rejected candidates.

This mechanism progressively suppresses repeated lures and forces the pooled search to converge on alternatives, effectively pruning a local attractor basin.


7. Optional: Contextual “Meaning Shift” Within Working Memory

To capture context-dependent remapping (the idea that the “same item” can shift meaning depending on the most recent update and current set), define a context-conditioned transform applied to each slot before pooling:W^t,k=Wt,k+Δ(Wt,k,qt),\hat W_{t,k} = W_{t,k} + \Delta(W_{t,k}, q_t),W^t,k​=Wt,k​+Δ(Wt,k​,qt​),

where Δ\DeltaΔ is a learned function (e.g., MLP). The pooled query is computed from W^t\hat W_tW^t​ rather than WtW_tWt​. This makes representational drift an explicit, measurable part of the model.


8. Training Objectives

The iterative working-memory module can be trained end-to-end inside larger models (transformer, recurrent agent, world model). The following losses are common and complementary.

8.1 Task loss (supervised or self-supervised)

If the system produces outputs yty_tyt​ (next token, next action, next-step label), train with a standard task loss:Ltask=tCE(yt,yt)orLtask=tytyt2.\mathcal{L}_{\text{task}} = \sum_t \text{CE}(y_t, y^*_t) \quad \text{or} \quad \mathcal{L}_{\text{task}} = \sum_t \|y_t – y^*_t\|^2.Ltask​=t∑​CE(yt​,yt∗​)orLtask​=t∑​∥yt​−yt∗​∥2.

8.2 Continuity regularization (optional)

To encourage a desired overlap regime, regularize the effective overlap O^t\hat O_tO^t​ toward target rrr:Loverlap=t(O^tr)2.\mathcal{L}_{\text{overlap}} = \sum_t (\hat O_t – r)^2.Loverlap​=t∑​(O^t​−r)2.

Here O^t\hat O_tO^t​ can be computed from keep masks (discrete) or from similarity of pooled summaries (soft).

8.3 World-model prediction (optional)

For predictive learning, minimize error in predicting next observation features ϕ(ot+1)\phi(o_{t+1})ϕ(ot+1​):Lpred=tϕ^t+1ϕ(ot+1)2.\mathcal{L}_{\text{pred}} = \sum_t \left\|\hat\phi_{t+1} – \phi(o_{t+1})\right\|^2.Lpred​=t∑​​ϕ^​t+1​−ϕ(ot+1​)​2.

This loss incentivizes the module to retain what is causally relevant across time while updating what changes.

8.4 Reinforcement learning (optional)

In RL settings, WtW_tWt​ conditions the policy π(atWt)\pi(a_t\mid W_t)π(at​∣Wt​). The working-memory module is trained jointly by actor–critic losses; overlap/continuity regularizers can be added to shape deliberation.


9. Developmental / Curriculum Scheduling of Retention

A key architectural hypothesis is that long-horizon coherence can be scaffolded by gradually increasing retention and persistence. Implement this via a schedule for rrr over training steps:r(step)=rmin+(rmaxrmin)σ ⁣(steps0s1),r(\text{step}) = r_{\min} + (r_{\max}-r_{\min})\cdot \sigma\!\left(\frac{\text{step}-s_0}{s_1}\right),r(step)=rmin​+(rmax​−rmin​)⋅σ(s1​step−s0​​),

where σ()\sigma(\cdot)σ(⋅) is the logistic function and s0,s1s_0,s_1s0​,s1​ control onset and steepness.

Early training (lower rrr) promotes exploration and rapid updating; later training (higher rrr) promotes compounding of intermediate results and stable long-horizon inference.


10. Key Knobs, Metrics, and Ablations

10.1 Knobs (interpretable control parameters)

  • Capacity: KKK (FoA size)
  • Retention / overlap: rrr (continuity)
  • Selection temperature: τ\tauτ (exploration vs convergence)
  • Inhibition: λ,κ,γ\lambda,\kappa,\gammaλ,κ,γ (escape from lures)
  • Pooling/keep functions: fpool,fkeepf_{\text{pool}}, f_{\text{keep}}fpool​,fkeep​ (learned control policy)

10.2 Metrics

  • Overlap / continuity: OtO_tOt​ or O^t\hat O_tO^t​
  • Reset frequency: incidence of low-overlap transitions
  • Lure perseverance: probability of re-selecting recently inhibited candidates
  • Long-horizon coherence: task-dependent (e.g., plan success, multi-step accuracy, discourse coherence)

10.3 Mechanism-proving ablations

Ablations that directly test whether each primitive matters:

  1. No overlap constraint: replace all KKK slots each step (set r=0r=0r=0)
  2. Overlap without pooled search: recruit using only the newest slot (single-cue)
  3. Pooled search without inhibition: remove hth_tht​ terms (λ=0\lambda=0λ=0)
  4. Fixed rrr vs curriculum r(⋅)r(\cdot)r(⋅): test developmental hypothesis
  5. Remove contextual remapping: set Δ()=0\Delta(\cdot)=0Δ(⋅)=0

These ablations tend to produce characteristic failures (derailment, lure trapping, weak bridging inference) that operationalize the theory’s claims.


11. Summary (the “kernel”)

The architecture reduces to a small set of commitments:

  1. A capacity-limited working set WtW_tWt​ evolves by an explicit operator

Wt+1=KeeprK(Wt)    Recruit(1r)K(Ctfpool(Wt)),W_{t+1} = \text{Keep}_{rK}(W_t) \;\cup\; \text{Recruit}_{(1-r)K}(C_t \mid f_{\text{pool}}(W_t)),Wt+1​=KeeprK​(Wt​)∪Recruit(1−r)K​(Ct​∣fpool​(Wt​)),

  1. New recruits are selected by pooled multi-cue search (softmax over candidate similarity to a pooled query),
  2. Search can be steered by iterative inhibition,
  3. Retention rrr is an interpretable continuity knob that can be trained and scheduled.

This formalization is simultaneously (i) a theory statement (what cognition is doing step-to-step) and (ii) a runnable AI design pattern (how to build an agent that maintains continuity and compounds intermediate results).

Posted in

Leave a comment