Iterated Insights

Ideas from Jared Edward Reser Ph.D.

Qualia as Transition Awareness: How Iterative Updating Becomes Experience

Abstract Qualia is often treated as a static property attached to an instantaneous neural or computational state: the redness of red, the painfulness of pain. Here I argue that this framing misidentifies the explanatory target. Drawing on the Iterative Updating model of working memory, I propose that a substantial portion of what we call qualia,…

Keep reading

Consciousness as Iteration Tracking: Experiencing the Iterative Updating of Working Memory

Abstract This article proposes a temporal and mechanistic model of consciousness centered on iterative updating and the system’s capacity to track that updating. I argue for three nested layers. First, iterative updating of working memory provides a continuity substrate because successive cognitive states overlap substantially, changing by incremental substitutions rather than full replacement. This overlap…

Keep reading

Does Superintelligence Need Psychotherapy? Diagnostics and Interventions for Self-Improving Agents

Abstract Agentic AI systems that operate continuously, retain persistent memory, and recursively modify their own policies or weights will face a distinctive problem: stability may become as important as raw intelligence. In humans, psychotherapy is a structured technology for detecting maladaptive patterns, reprocessing salient experience, and integrating change into a more coherent mode of functioning.…

Keep reading

Why Transformers Approximate Continuity, Why We Keep Building Prompt Workarounds, and What an Explicit Overlap Substrate Would Change

Abstract This article argues that “continuity of thought” is best understood as the phenomenological signature of a deeper computational requirement: stateful iteration. Any system that executes algorithms across time needs a substrate that preserves intermediate variables long enough to be updated, otherwise it can only recompute from scratch. Using this lens, I propose a simple…

Keep reading

Something went wrong. Please refresh the page and/or try again.

I. Introduction: The Problem of Continual Learning

One of the central ambitions of artificial intelligence research today is continual learning: the ability for a system to keep learning indefinitely after pretraining, without catastrophic forgetting, brittleness, or the need for full retraining. Despite decades of work, this goal remains elusive. Most modern AI systems are either highly capable but static, or adaptive but unstable. They learn impressively during training, yet struggle to integrate new knowledge once deployed.

The dominant failure modes are well known. When models continue to update their parameters, new learning often overwrites old knowledge. When updates are restricted to prevent forgetting, learning stalls. The result is a persistent tension between plasticity and stability, with no generally accepted resolution.

This essay argues that the difficulty of continual learning is not merely a technical problem, but a conceptual one. Most approaches implicitly treat learning as accumulation—adding new representations, protecting old ones, or balancing updates between them. Biological intelligence, by contrast, does not scale indefinitely by accumulation. It scales by restructuring.

The central claim of this essay is that continual learning becomes tractable when learning is reframed as iterative compression rather than indefinite accumulation. Systems that can repeatedly re-encode experience into simpler, more invariant representations can integrate new knowledge without overwriting the old, because old knowledge has already been abstracted into forms that are robust to change.

The theoretical basis for iterative compression—derived from working memory dynamics, attention, and long-term memory plasticity—is developed in detail in a companion essay:

The present essay builds on that foundation and focuses specifically on the implications for continual learning in artificial systems.

A critical but easily overlooked point is that iterative compression is not an abstract post-hoc process applied to stored representations; it is actively driven by the moment-to-moment mechanics of iterative updating in working memory and attention. In both humans and artificial systems during training and inference, each update of working memory operates over pre-existing conceptual sets already associated in long-term memory. Attention does not sample representations arbitrarily. It selectively stabilizes some elements, suppresses others, and recruits nearby associations through spreading activation. As this process repeats, slightly different subsets of the same underlying conceptual neighborhood are brought into co-activation, compared against goals and error signals, and either retained or pruned. Over time, this iterative cycling fine-tunes the boundaries of these conceptual sets: unstable elements are progressively excluded, redundant distinctions collapse, and consistently co-active elements become tightly bound. Iterative compression therefore emerges from the dynamics of attention-driven iteration itself. The system is not compressing a static representation, but repeatedly revising which elements belong together, gradually reshaping long-term memory so that future iterations recruit cleaner, more invariant sets with fewer intermediate steps. In this way, the mechanics of iterative updating in working memory are the causal engine that sculpts compression, rather than a process that merely follows it. You can read all about this on my website aithought.com

II. What Current Continual Learning Approaches Get Right—and Wrong

Existing approaches to continual learning are not misguided. Many correctly identify the symptoms of the problem and propose partial remedies.

Replay-based methods attempt to preserve old knowledge by revisiting past experiences, either by storing data directly or by generating approximate reconstructions. Regularization-based methods penalize changes to parameters deemed important for previous tasks. Architectural approaches introduce modularity, expandable networks, or task-specific components to isolate interference.

These methods address real failure modes, and in constrained settings they can be effective. However, they share a common limitation: they treat knowledge as something to be protected or preserved in its existing form.

Replay alone preserves details without simplifying them. Regularization freezes structure without improving it. Modular architectures avoid interference by separation, but at the cost of fragmentation and unbounded growth. In all cases, the system retains increasingly complex internal representations that must coexist indefinitely.

What is missing is a mechanism for progressive simplification. Human learners do not indefinitely preserve the full structure of early representations. Instead, early, detailed representations are gradually replaced by abstractions that subsume them. Continual learning systems fail not because they forget too easily, but because they fail to compress.

III. Iterative Compression as a Process, Not a Property

Compression in machine learning is often treated as a static property: a trained model is said to “compress” data if it uses fewer parameters or lower-dimensional representations. Iterative compression, by contrast, is a process that unfolds over time.

Iterative compression refers to the repeated re-encoding of experience such that representations become simpler while preserving functional performance. Each compression pass removes redundancy, discards unstable detail, and retains invariant structure. Crucially, compression is not performed once; it is revisited repeatedly as new information arrives and as the system’s internal model evolves.

In biological cognition, this process is driven by iterative updating in working memory, guided by attention and error signals, and consolidated into long-term memory through plasticity. Over time, multi-step reasoning paths collapse into direct associations. Rich episodic traces give way to abstract schemas. Learning proceeds not by storing more, but by needing less.

This distinction is critical for continual learning. Systems that only accumulate representations must either protect everything or risk forgetting. Systems that iteratively compress can integrate new information by rewriting old knowledge into simpler forms that remain compatible with future learning.

You can find my writings about this at aithought.com and here are the published articles I’ve written on the subject:

A Cognitive Architecture for Machine Consciousness and Artificial Superintelligence: Updating Working Memory Iteratively, 2022 arXiv:2203.17255 

And 

Reser JE. 2016. Incremental change in the set of coactive cortical assemblies enables mental continuity. Physiology and Behavior. 167: 222-237.

IV. How Iterative Compression Addresses Core Continual Learning Failures

Iterative compression offers a principled solution to several long-standing problems in continual learning.

Catastrophic forgetting arises when new updates overwrite fragile, task-specific representations. Iterative compression reduces this fragility by promoting only stable invariants into long-term structure. What is protected is not raw experience, but compressed representations that already summarize many past experiences. New learning is expressed through these abstractions rather than in conflict with them.

Generalization and transfer improve naturally under compression. Representations that survive repeated re-encoding are, by definition, those that apply across contexts. As a result, compressed representations support reuse and transfer without requiring explicit task boundaries.

The stability–plasticity dilemma is reframed rather than balanced. Plasticity operates on rich, high-dimensional representations early in learning. Stability applies only after compression has identified what deserves protection. Stability is therefore not imposed globally; it is earned locally.

Finally, iterative compression suggests new metrics for learning progress. Instead of measuring only task accuracy, one can track reductions in representational complexity, shorter inference paths, or the stability of internal attractors over time. Learning progress is reflected not just in what the system can do, but in how simply it can do it.

V. Architectural Consequences for Continual Learning Systems

If iterative compression is essential for continual learning, then certain architectural features become necessary rather than optional.

First, systems require a persistent working memory loop—a capacity-limited, temporally continuous workspace in which representations can be repeatedly re-evaluated and revised. Without such a loop, learning updates remain global and indiscriminate.

Second, long-term memory must be rewritable, not merely append-only or frozen. Continual learning demands that older representations be reformulated in light of new experience, rather than preserved indefinitely in their original form.

Third, replay must function as reconstruction, not rehearsal. Replayed experiences should be reinterpreted under the current model and reconsolidated in compressed form. Simply repeating stored patterns preserves complexity without yielding abstraction.

Finally, systems need compression scheduling: alternating phases of new learning and consolidation, analogous to offline learning in biological systems. Continual learning is not continuous gradient descent; it is a rhythm of acquisition and reorganization.

Together, these requirements point toward a class of architectures that learn indefinitely not by growing without bound, but by continually simplifying themselves while preserving what works.

VI. Iterative Compression in Relation to Existing Learning Paradigms

Iterative compression does not replace existing paradigms in continual learning; rather, it clarifies what they are missing and how they might be unified.

In meta-learning, systems learn to adapt quickly by discovering higher-order learning rules. Iterative compression can be understood as a complementary process operating on a longer timescale. Meta-learning accelerates acquisition; iterative compression stabilizes what is acquired by distilling it into invariant structure. Without compression, meta-learned flexibility risks accumulating fragile representations.

In representation learning, bottlenecks and regularization are often used to encourage abstraction. However, these mechanisms are typically static, applied during a single training phase. Iterative compression generalizes this idea temporally: bottlenecks are revisited repeatedly, and representations are forced to survive under successive reinterpretations rather than a single optimization objective.

Replay-based systems come closest to capturing the spirit of iterative compression, but usually fall short in execution. Replay that merely preserves old experiences defends the past without improving it. Iterative compression requires replay to function as reinterpretation, where past experiences are reconstructed under the current model and rewritten in simpler form. Without this rewriting step, replay stabilizes complexity instead of reducing it.

Seen in this light, iterative compression provides a missing throughline connecting meta-learning, representation learning, and replay—while explaining why none of them alone has solved continual learning.

VII. Predictions and Empirical Signatures

If iterative compression is a necessary condition for robust continual learning, then it makes several concrete, testable predictions.

First, systems capable of indefinite learning should exhibit declining representational complexity over time, even as task performance remains stable or improves. Complexity may be measured through dimensionality, description length, inference depth, or internal path length.

Second, continual learning systems should benefit from explicit offline phases during which no new data is introduced. Performance gains following such phases would indicate successful reorganization rather than mere accumulation.

Third, systems that compress effectively should display graceful degradation under distributional shift. Because compressed representations capture invariants rather than surface detail, they should fail conservatively rather than catastrophically.

Finally, the absence of compression should predict long-term brittleness. Systems that continually add structure without simplification should show increasing interference, longer inference paths, and greater susceptibility to edge cases as learning progresses.

These predictions distinguish iterative compression from vague appeals to “better regularization” and place it squarely in the domain of falsifiable theory.

VIII. Broader Implications Beyond Artificial Intelligence

Although this essay is framed around artificial systems, the implications of iterative compression extend beyond AI.

In cognitive science, the framework explains why expertise is associated with both speed and simplicity. Experts do not possess more detailed representations than novices; they possess more compressed ones. What looks like intuition is often the result of extensive prior compression.

In education, the framework clarifies why explanation, teaching, and rewriting are such powerful learning tools. These activities force representations to survive reformulation under new constraints, accelerating compression.

In neuroscience, iterative compression aligns naturally with memory reconsolidation, sleep-dependent learning, and the observed shift from hippocampal to cortical representations over time. These phenomena are difficult to explain under pure accumulation models, but follow naturally from a compression-based account.

More broadly, iterative compression reframes cognitive limitations—finite working memory, forgetting, attentional bottlenecks—not as flaws, but as drivers of intelligence. Without pressure to simplify, learning systems would accumulate endlessly and fail to generalize.

IX. Why Continual Learning Has No Shortcut

One reason continual learning has resisted solution is that it cannot be solved by a single mechanism. There is no regularizer, memory buffer, or architectural tweak that can substitute for repeated reorganization over time.

Iterative compression is slow, conservative, and retrospective. It depends on hindsight. It requires revisiting the past and rewriting it. These properties make it difficult to engineer and easy to overlook—but they are precisely what allow biological systems to learn indefinitely without collapse.

Attempts to bypass compression by freezing models, isolating modules, or endlessly replaying experiences treat the symptoms of continual learning failure without addressing its cause. They preserve what exists instead of asking what can now be safely removed.

Continual learning, in the strong sense, demands systems that are willing to forget details in order to remember structure.

X. Conclusion: Continual Learning as Continuous Re-Compression

The central argument of this essay is simple but far-reaching: continual learning is not continuous accumulation; it is continuous re-compression.

Artificial systems struggle to learn indefinitely because they lack mechanisms for progressively simplifying their own representations while preserving what works. Biological intelligence succeeds not by storing more and more detail, but by repeatedly rewriting experience into increasingly compact, invariant forms.

Iterative compression provides a unifying principle that explains why current approaches to continual learning fall short, how they can be improved, and what architectural features are truly required. It reframes the stability–plasticity dilemma, clarifies the role of replay, and offers new metrics for evaluating learning progress.

Most importantly, it restores time to the center of intelligence. Learning is not a moment, but a history. Systems that cannot revisit and revise their own past will always remain bounded. Systems that can iteratively compress what they know may, for the first time, be able to learn without end.

Posted in

Leave a comment