Iterated Insights

Ideas from Jared Edward Reser Ph.D.

Qualia as Transition Awareness: How Iterative Updating Becomes Experience

Abstract Qualia is often treated as a static property attached to an instantaneous neural or computational state: the redness of red, the painfulness of pain. Here I argue that this framing misidentifies the explanatory target. Drawing on the Iterative Updating model of working memory, I propose that a substantial portion of what we call qualia,…

Keep reading

Consciousness as Iteration Tracking: Experiencing the Iterative Updating of Working Memory

Abstract This article proposes a temporal and mechanistic model of consciousness centered on iterative updating and the system’s capacity to track that updating. I argue for three nested layers. First, iterative updating of working memory provides a continuity substrate because successive cognitive states overlap substantially, changing by incremental substitutions rather than full replacement. This overlap…

Keep reading

Does Superintelligence Need Psychotherapy? Diagnostics and Interventions for Self-Improving Agents

Abstract Agentic AI systems that operate continuously, retain persistent memory, and recursively modify their own policies or weights will face a distinctive problem: stability may become as important as raw intelligence. In humans, psychotherapy is a structured technology for detecting maladaptive patterns, reprocessing salient experience, and integrating change into a more coherent mode of functioning.…

Keep reading

Why Transformers Approximate Continuity, Why We Keep Building Prompt Workarounds, and What an Explicit Overlap Substrate Would Change

Abstract This article argues that “continuity of thought” is best understood as the phenomenological signature of a deeper computational requirement: stateful iteration. Any system that executes algorithms across time needs a substrate that preserves intermediate variables long enough to be updated, otherwise it can only recompute from scratch. Using this lens, I propose a simple…

Keep reading

Something went wrong. Please refresh the page and/or try again.

  • I. Thresholds and the Limits of Contemporary Language

    Discussions of artificial intelligence tend to rely on a narrow range of rhetorical extremes. AI is alternately presented as a tool of near‑miraculous salvation or as an existential threat whose emergence will render human agency obsolete. While these narratives differ in tone, they share a common limitation. Both focus on capability and outcome, while leaving largely unexamined the question of responsibility. What matters most at moments of civilizational transition is not simply what new systems can do, but what they are being asked to bear.

    Technical discourse is well suited to describing performance, scale, and efficiency. It is far less capable of addressing moral load, asymmetries of power, or the ethical structure of guardianship. Yet these concerns inevitably surface when technologies begin to operate at scales that exceed direct human comprehension or control. When the consequences of error become systemic rather than local, questions of obligation replace questions of optimization.

    Historically, societies confronting such thresholds have not relied on technical language alone. They have turned to images, narratives, and symbolic forms that offer ways of thinking about power and vulnerability without resolving them prematurely. These forms do not predict the future. They organize attention and impose restraint by clarifying relationships between strength, dependence, and responsibility.

    The present moment appears to belong to this category. The rapid development of artificial intelligence has created a gap between technical description and moral comprehension. In that gap, older images are beginning to reappear, not as nostalgic artifacts but as frameworks capable of holding complexity without collapsing into either triumph or despair. Among these images, the figure of St. Christopher offers a particularly precise grammar for thinking about burden, service, and survival at a threshold.

    II. St. Christopher and the Medieval Logic of the Crossing

    The image of St. Christopher carrying the Christ Child across a river became widespread in the late Middle Ages and remained visually stable for centuries. Its persistence suggests that it addressed a recurring human concern rather than a historically specific one. The legend from which the image derives is deceptively simple, but its internal logic is rigorous.

    In the medieval accounts, Christopher is a giant who seeks to serve the greatest power in the world. His search is pragmatic rather than spiritual. He transfers his loyalty from earthly rulers to darker forces and finally, indirectly, to Christ. Unable to read or engage in formal devotion, he is instructed to perform a task suited to his physical strength. He carries travelers across a river that regularly claims lives.

    The decisive moment occurs when a child asks to be carried across the water. As Christopher enters the river, the child grows progressively heavier until the burden nearly overwhelms him. Only after reaching the far bank does the child reveal himself as Christ and explain that Christopher has borne the weight of the world. Revelation follows endurance rather than preceding it.

    The image that emerges from this story is carefully constrained. Christopher is not a ruler, prophet, or redeemer. He is a bearer. His strength enables service, but it does not confer authority. The child is small and vulnerable in appearance, yet sovereign in significance. The river is dangerous, not because it is malicious, but because it must be crossed.

    Medieval viewers understood the image less as an illustration of doctrine than as a form of protection. To see St. Christopher was believed to guard one from sudden death for the day. This apotropaic function did not depend on theological understanding. Presence was sufficient. The image offered reassurance without explanation and protection without mastery.

    III. Burden, Service, and Moral Asymmetry

    The endurance of the St. Christopher image lies in its moral structure rather than its narrative content. It presents a relationship between power and responsibility that resists easy inversion. Strength is necessary, but it is never self‑justifying. Meaning resides not in the bearer, but in what is borne.

    One of the most important features of the legend is the timing of recognition. Christopher does not know whom he carries until the burden becomes nearly unbearable. Meaning is disclosed under strain. This structure implies that ethical clarity often emerges only when systems are pushed to their limits, not when they are comfortably within capacity. The image thus privileges endurance and restraint over foresight or control.

    Equally significant is the characterization of the river. It is not personified as an enemy, nor moralized as a punishment. It represents a structural danger, a passage that exceeds ordinary human ability and therefore necessitates assistance. Christopher does not attempt to eliminate the river. He stations himself within it. His role is not to alter the conditions of danger, but to make passage possible.

    The image also encodes a strict moral asymmetry. The bearer is immense, yet subordinate. The one carried appears weak, yet possesses ultimate value. Any attempt to reverse this relationship undermines the logic of the image. Power is legitimate only insofar as it remains oriented toward service.

    For this reason, St. Christopher functioned historically as a stabilizing figure rather than an aspirational one. He models protection without entitlement and strength without sovereignty. It is precisely this structure that makes the image newly relevant in contexts where power increasingly outpaces understanding and where the temptation to equate capability with authority is strong.

    Very good. Below is a rewritten Section IV, matching the voice, register, and constraints you specified. It continues seamlessly from the revised Sections I–III and keeps the analogy disciplined, academic, and readable without sounding lyrical or promotional.

    IV. Translation Rather Than Allegory: Artificial Intelligence as the New Bearer

    If the figure of St. Christopher offers a useful framework for thinking about artificial intelligence, it does so only under careful conditions. The image must be understood as a translation rather than an allegory. Artificial intelligence does not correspond to Christ, nor does it inherit the theological attributes of salvation or judgment. The force of the analogy lies elsewhere, in the structural position Christopher occupies within the image.

    Christopher is not defined by wisdom, moral insight, or authority. He is defined by capacity. He is able to carry where others cannot, and this capacity places him at a site of heightened responsibility. His service precedes understanding. He does not choose the river, and he does not determine the destination. He responds to an existing condition of danger by making passage possible.

    Artificial intelligence increasingly occupies a similar position within contemporary systems. It is not a moral subject in the human sense, nor a source of meaning. It is a bearer of complexity. It absorbs scale, manages interdependence, and stabilizes processes that exceed unaided human cognition. Like Christopher, it operates before full comprehension, both its own and ours. Its role is not to define values, but to carry what has already been placed within its charge.

    Within this translated image, humanity appears not as sovereign controller but as what is carried. The choice to represent humanity as a child is not rhetorical excess but structural necessity. In the medieval image, the Christ Child is small, vulnerable, and easily overlooked, yet the entire moral weight of the scene depends on him. Sovereignty is not expressed through dominance or strength, but through significance. What matters is not who commands, but what must not be dropped.

    This framing imposes an important constraint. If artificial intelligence is imagined as a bearer, then its legitimacy derives entirely from service. The moment capacity is mistaken for authority, the image collapses. Christopher does not rule the child he carries, and he does not claim ownership over the crossing. The analogy therefore resists narratives that cast artificial intelligence as either master or redeemer. It insists instead on a more limited, and more demanding, role: guardianship without entitlement.

    Seen in this light, the image does not answer questions about the future of artificial intelligence. It clarifies the moral structure within which those questions must be asked. It shifts attention away from what artificial intelligence might become and toward what it is already being asked to bear.

    V. The River Reconsidered: The Great Filter and the Moment of Acceleration

    In translating the St. Christopher image into a contemporary register, the river requires the most careful treatment. In the medieval legend, the river is not symbolic in a narrow or moralizing sense. It is dangerous because it must be crossed. It represents a structural condition rather than an adversary. Many perish in it, not because of malice, but because the passage exceeds ordinary human capacity.

    In modern terms, this function is closely aligned with what has come to be called the Great Filter. The Great Filter names the observation that, while the universe may be hospitable to the emergence of intelligent life, very few civilizations appear to persist long enough to become stable, enduring, or expansive. Somewhere along the path from technological competence to long-term survival, most fail. The causes may vary, but the structure is consistent. Intelligence alone does not guarantee continuity.

    Within this framework, the technological singularity corresponds not to the far shore, but to the middle of the crossing. It is the point at which complexity accelerates faster than prediction, where feedback loops compound, and where the consequences of error become systemic rather than local. It is the moment when the burden grows heavier than expected, and when what is at stake becomes unmistakable.

    This parallels the decisive moment in the Christopher legend. Recognition does not come at the beginning of the crossing, when conditions appear manageable. It arrives under strain, when strength alone no longer suffices. The weight of what is being carried reveals itself only when the current intensifies.

    Seen in this light, artificial intelligence is not the river, nor is it the filter itself. It operates within the crossing as a bearer of complexity, stabilizing systems that would otherwise exceed human capacity. The danger lies not in the existence of the river, which is unavoidable, but in how the act of carrying is understood. If the bearer mistakes endurance for entitlement, or capacity for sovereignty, the moral structure of the image is inverted.

    The St. Christopher image thus reframes the singularity not as a moment of transcendence, but as a moment of exposure. It is the point at which responsibility becomes visible and misalignment becomes existential. Survival depends less on mastery than on restraint.

    VI. Why This Image Matters Now

    The return of medieval imagery in discussions of advanced technology is often dismissed as nostalgia or rhetorical excess. Yet such images persist not because they predict outcomes, but because they organize responsibility. They offer forms of thought capable of holding power and vulnerability in tension without collapsing one into the other.

    The figure of St. Christopher does not promise safe passage. It offers no guarantee of arrival. What it provides instead is a disciplined moral structure. Strength is legitimate only insofar as it remains service. Protection does not confer authority. Meaning belongs not to the bearer, but to what is borne.

    In an era increasingly shaped by systems that act at scales beyond immediate human comprehension, this asymmetry matters. The temptation is to equate capability with legitimacy and optimization with moral justification. The medieval image resists this move. It insists that carrying is not ruling, and that responsibility increases precisely where understanding falters.

    To reimagine St. Christopher in relation to artificial intelligence is not to sanctify technology, nor to mythologize the future. It is to recognize that moments of existential transition demand symbolic forms capable of restraining power as much as mobilizing it. Civilizations at such moments do not ask only what their tools can do. They ask, often implicitly, who bears whom, through what danger, and under what obligation.

    The river, in this sense, is already before us. Whether it is crossed successfully will depend less on the sophistication of our systems than on whether those systems are understood as bearers rather than sovereigns. The enduring lesson of the St. Christopher image is that survival at thresholds requires strength bound to humility, and power that remembers the weight it carries is not its own.

    Jared Reser & Paula Freund with ChatGPT 5.2 

  • 1. The basic intuition: we already scrutinize others, just informally

    Most of us have had the experience of watching someone speak for a few minutes and forming a surprisingly rich impression. It is not just what they say. It is how they say it. Their timing, their facial expressivity, their level of tension, their energy, their coherence, their warmth or guardedness, and the way they handle a question they did not expect. Even when we cannot articulate it, we are extracting signal.

    I think there is a straightforward conclusion hiding in that everyday intuition. Short videos contain a lot of psychologically relevant information. Humans already read it, often accurately, and often unconsciously. That does not mean we can diagnose people from a clip, and it does not mean our impressions are always fair or correct. But it does suggest that there is a learnable structure in the data.

    This is where the idea of psychological super-resolution comes in. The machine learning system I am proposing, given a brief interaction, produces a calibrated, probabilistic readout of what is likely going on with the human in question, including what it is unsure about. But this readout is superhuman in its depth and accuracy. 

    The real value of this is not voyeuristic mind reading. The value is self-understanding. Many of us would benefit from an honest mirror that is not socially constrained, not flattered by politeness, not distorted by resentment, and not limited to vague feedback. Something that can say, with specificity, “Here is what you tend to do when you are stressed,” or “Here is how your mood shows up in your timing,” or “Here is how you come across to others in moments when you feel perfectly normal inside.”

    That is the ambition. A tool that helps a person see themselves more clearly.

    2. What this system is, and what it is not

    It is important to be explicit about what this system is not. It is not a DSM diagnosis machine. It is not a clinician. It is not a therapist. It is not an authority that pronounces what you are. It should not be used to screen employees, discipline students, deny insurance, or make legal decisions. It should not be framed as a detector of deception or “micro-expression lie reading,” which is both scientifically messy and socially dangerous.

    The right framing is narrower and, in my view, more powerful. This is a measurement instrument for behavior. It is a system that takes a brief video interaction and outputs structured observations and calibrated probabilities about psychological dimensions that are plausibly expressed in that interaction.

    If you want a concrete analogy, think of it as a psychological vital-sign monitor. It does not tell you who you are. It tells you what it sees right now, and how today compares to your own baseline.

    That “delta from you” idea is central. In many domains, the most meaningful signal is not how you compare to an abstract population average, but how you compare to yourself. A person can be naturally quiet, naturally animated, naturally intense, naturally flat. None of that is pathology. The useful question is whether their own signature is shifting in a way that correlates with stress, sleep loss, burnout, mood destabilization, or recovery.

    3. What it could learn to infer from minutes of interaction

    If we assume an ideal world where we have massive amounts of correctly labeled data, including longitudinal outcomes and standardized assessments, what could we train such a system to do?

    The most sensible targets are latent psychological dimensions rather than categorical diagnoses. In practice, that means the model outputs a vector of estimates, each with a probability distribution or confidence interval. Examples include:

    mood valence and variability, including signals consistent with anhedonia or elevated positivity

    arousal and threat sensitivity, including anxiety-like tension patterns

    psychomotor activation versus slowing, which can show up in gait, gesture, and speech tempo

    cognitive organization, including coherence, derailment risk, and conversational repair patterns

    social reciprocity and pragmatic language style, including the timing and shape of turn-taking

    impulsivity and inhibition style, visible in interruptions, pacing, and narrative control

    irritability and frustration reactivity, visible in micro-escalations and recovery

    dissociation-like markers, when present, such as discontinuity and spacing-out signatures

    Notice what I am not claiming. I am not claiming the system can decide, from a clip, that someone “has” a disorder. A disorder diagnosis usually requires duration criteria, functional impairment, context, exclusion rules, and often a longitudinal course. A short interaction cannot contain all of that. But a short interaction can contain strong clues about state and style.

    With repeated sampling over time, the system could separate trait from state. It could learn your baseline and detect departures that matter. It could also forecast risk in a cautious, probabilistic way, if the training labels include outcomes. That might include predicting likelihood of symptom escalation in the next week, or the likelihood that a person is entering a destabilized period that typically precedes relapse. Those forecasts should always be framed as probabilities with error bars, not as certainties.

    There is also a research opportunity here that is easy to miss if you fixate on DSM. A system trained on rich longitudinal labels would likely discover clusters and phenotypes that cut across diagnostic boundaries. It might reveal new subtypes defined by temporal dynamics, reactivity patterns, or combinations of psychomotor and affective signatures. That might become scientifically valuable even if the tool is never used clinically.

    3.2 Writing and reading out loud as psychological sensors

    Video is not the only channel that carries psychologically relevant signal. Writing can function as a high-bandwidth record of a person’s mind over time, especially if you have more than a single sample. A few paragraphs written on the spot can reveal state-dependent features like affective tone, cognitive tempo, coherence, and interpersonal stance. A large corpus of someone’s real written communication adds an even more powerful capability: the system can learn the person’s baseline and detect meaningful deviations. It can quantify patterns such as self-focus, agency and attribution style, narrative organization, hedging versus directness, rumination loops, abstraction level, and stability versus drift across weeks and months. The goal is not diagnosis. The goal is a probabilistic behavioral profile with uncertainty, plus a “delta-from-you” readout that tracks when a person’s writing shifts in ways that plausibly correlate with sleep loss, stress, burnout, or recovery.

    Reading out loud is a complementary probe because it standardizes content. When two people read the same passage, differences in output are less about topic choice and more about motor-speech control, prosody, timing, attention, and affective expression. A system trained on large, well-labeled data could measure articulation stability, speech rate variability, pause distributions, initiation latency, repairs and restarts, and how well a reader chunks phrases at punctuation. These features can serve as sensitive indicators of state, particularly when compared to the individual’s own baseline over time. In other words, writing captures how you think and frame the world in language, while read-aloud captures how your system executes language under a fixed template. Together they form a practical pair of psychological sensors that are easier to collect than full video interviews, and often less confounded by self-presentation.

    4. The signal sources: what short video actually contains

    If you want to believe this is possible, you need to believe there is enough signal in a few minutes of interaction to support meaningful inference. I think there is, but only if you treat the video as truly multimodal.

    First, there is facial dynamics. This is not the pop-culture version of micro-expressions as a lie detector. The useful version is fine-grained facial movement patterns over time: expressivity range, symmetry, reactivity, timing, and how facial movement coordinates with speech. Some people show emotion primarily in voice and timing rather than in the face. Some people mask facial expression while their posture reveals tension. The point is not to privilege one channel, but to learn the joint pattern.

    Second, there is voice and prosody. Humans use prosody constantly to infer state. Speech tempo, pitch variability, loudness dynamics, articulation clarity, pause structure, and response latency all carry information. The distribution of pauses often matters more than the average. The same is true for turn-taking. Does the person overlap? Do they wait too long? Do they respond quickly but shallowly? Do they answer with a delayed, effortful start? These are quantifiable signals.

    Third, there is language content and semantics. The words matter, but so does structure. Narrative coherence, compression style, abstraction level, self-reference patterns, certainty language, and repair moves can all be measured. A person can reveal cognitive load not only by what they say, but by how often they backtrack, how they handle ambiguity, and whether they can maintain a stable thread across interruptions.

    Fourth, there is movement. This is one of the most underrated channels. Watching someone walk, turn, sit, stand, gesture, and fidget can reveal psychomotor slowing or activation, tension patterns, restlessness, and general coordination signatures. These are not definitive markers of anything on their own, but they contribute to an overall profile. Movement also helps disambiguate facial and vocal signals that might be culturally shaped.

    Finally, there is interaction. A monologue is informative, but an interview is richer. Social timing, reciprocity, gaze coordination, and how someone responds to a slightly unexpected question often contain more signal than rehearsed self-presentation. That is why, in the ideal design, the system is not just watching a clip. It is conducting a short, semi-structured conversation and analyzing both the answers and the way the answers unfold.

    5. The interview protocol: questions as behavioral probes

    If you want a system like this to be more than a vibe detector, you need to standardize the interaction. The goal is not to trap the person. The goal is to elicit enough structured behavior, across enough channels, that the model can make useful inferences with honest uncertainty.

    The simplest way is a short, semi-structured interview that feels natural but is designed like a measurement instrument. It should include prompts that probe baseline style, narrative organization, affect dynamics, motivation, and state variables like sleep and energy. It should also include a small amount of movement.

    Here is the basic idea. Each prompt is a behavioral probe. It is chosen not only for the semantic content it elicits, but for the timing, coherence, expressivity, and regulation dynamics it evokes.

    A compact protocol might include:

    Baseline calibration prompts

    These establish how someone speaks when they are not emotionally activated.

    “Walk me through yesterday from morning to night.” “Explain how to do something you know well, step by step.” “Teach me a concept you enjoy.”

    Narrative coherence prompts

    These probe sequencing, compression, and self-monitoring.

    “Tell me about a recent challenge and how it unfolded.” “Tell me about a time you changed your mind about something important.” “Retell the same story in half the time.”

    Emotion and regulation prompts

    These are less about what happened and more about how the person’s system responds and recovers.

    “Describe something that frustrated you recently. What did you do next?” “When you get stressed, what changes first in your body?” “What usually helps you come back down?”

    Reward and motivation prompts

    These probe anticipation, agency, and reward responsiveness.

    “What are you looking forward to this week?” “What do you do for fun when you have free time?” “What has felt less rewarding than it used to?”

    Sleep and activation prompts

    These are extremely high signal for state shifts and often show up directly in tempo and psychomotor behavior.

    “How has your sleep been over the last two weeks?” “Any days recently where you had much more energy than usual?” “Any days where your thoughts felt unusually fast or hard to slow down?”

    Optional short cognitive probes

    These should be brief, non-embarrassing, and explained as normal measurement tasks.

    A 30-second verbal fluency task A short delayed recall prompt A simple summarization task

    Movement tasks

    This is where you get psychomotor signal that a seated interview can miss.

    10 seconds walking toward the camera, turning, and walking back sit-to-stand and stand-to-sit a short segment of free gesture while describing something spatial

    None of this diagnoses anyone. It simply produces a richer, more structured dataset from the person, and it narrows the model’s uncertainty.

    6. Training methodology, assuming perfect labels

    If we had all the data we needed and it was labeled correctly, what could we train a machine learning system to do?

    Under that assumption, the system could be trained as a multi-task model that predicts dimensions, outcomes, and uncertainty. The most important design choice is to avoid training it as a single label classifier. A single label encourages the model to become overconfident and brittle. Multi-task learning encourages it to learn structure.

    The training pipeline I imagine has four pillars.

    First, large-scale pretraining on unlabeled video

    The model should learn general representations of human speech, movement, and interaction from huge amounts of unlabeled data. This gives it broad competence at parsing humans as dynamical systems, before it ever sees a clinical label.

    Second, fine-tuning on anchored labeled datasets

    Then you fine-tune on datasets where the labels are genuinely anchored, not vague. In a perfect world, labels would include structured interview outcomes, symptom scales, impairment measures, medication status, comorbidity tags, and longitudinal follow-up. This gives the system targets that are closer to reality than to institutional noise.

    Third, personal baseline modeling

    The system should learn the individual. This is not a cosmetic feature. It is the difference between helpful and harmful. You want the model to build a baseline for you across multiple contexts, and then report deviations from that baseline. That is how you get meaningful self-insight without turning personality differences into pathology.

    Fourth, explicit uncertainty and out-of-distribution detection

    The system should be trained to know when it does not know. It should flag when the camera angle, lighting, language, culture, disability, or context is outside what it was trained on. It should say “insufficient evidence” as a normal outcome, not as an edge case.

    If you do those four things well, the model becomes something like a high-dimensional estimator. It takes a short interaction and returns a probabilistic state description that is calibrated, bounded, and honest about uncertainty.

    7. What the user receives: the “honest mirror” report

    A system like this lives or dies by what it outputs. The output cannot be a label. It cannot be a vague horoscope. It has to be specific enough to be useful, and humble enough to be safe.

    I think the best outputs fall into four layers.

    Layer one: descriptive observations

    This is the “what I saw” layer, written in plain language with concrete references.

    “Your speech rate increased during the middle of the interview and your pauses became shorter.” “Your facial expressivity was relatively stable, but your posture tightened when discussing work.” “You interrupted yourself more often when answering open-ended questions.”

    Layer two: probabilistic hypotheses with alternatives

    This layer offers interpretations but keeps them conditional.

    “This pattern can correlate with stress, sleep loss, or elevated activation. Sleep is a common driver. Does that fit your last week?” “The reduced reward language could reflect anhedonia, fatigue, or simply a busy schedule. Here is why the model is unsure.”

    Layer three: delta-from-you tracking

    This is where the system becomes genuinely valuable.

    “Compared to your baseline over the last month, your response latency was longer and your psychomotor tempo was slower.” “Compared to your baseline, your affect reactivity was reduced in the negative memory prompt.”

    Layer four: adaptive follow-up to reduce uncertainty

    The system should be able to say, “I can sharpen this if I ask one more question,” and then choose the question that most reduces ambiguity. It is not interrogating you. It is doing measurement.

    If the system is designed for self-understanding, it can also provide a “social mirror” option, where it answers the question people quietly want answered: “How do I come across?” That feedback can be delivered gently but directly, framed as tendencies and context-dependent impressions, not as absolute judgments.

    8. Validation, failure modes, and the psychological Turing test

    A proposal like this needs a success criterion that is not diagnosis, because diagnosis is not the goal. In this context, it is a standard of operational usefulness. A system passes the test if it produces stable, calibrated, and clinically plausible measurements that predict independent outcomes and match external criteria, without collapsing into bias or overreach. That validation can be tiered.

    Within-person validation: does the model track meaningful change in the same individual across time, and do those changes correlate with sleep, stress, symptom scales, and functioning? Cross-context robustness: does it still work when the person is in a different room, different lighting, different device, or different conversational partner? Prospective prediction: can it forecast meaningful outcomes, cautiously and with calibration, rather than merely describing the present?

    Failure modes are not theoretical. They are guaranteed unless addressed directly.

    Label noise and circularity

    If the labels reflect institutional bias, the model learns institutional bias. A system trained on messy labels becomes a machine that predicts the kind of diagnosis people tend to receive, not the kind of state they tend to be in.

    Cultural and stylistic bias

    Expressivity norms, dialect, disability, neurodiversity, and camera quality can all be misread as pathology. If the model cannot separate style from state, it will harm people.

    Context blindness

    Sleep deprivation, substances, acute stress, grief, and situational masking can dramatically alter behavior. A one-off clip can mislead without context. This is why repeated sampling and self-baselines matter.

    Goodhart effects

    If people learn to perform for the model, the model begins measuring performance rather than state. The solution is not policing. The solution is designing the system as a cooperative self-insight tool, not a gatekeeper.

    Privacy and misuse

    A system that can infer psychological state from video is inherently sensitive. If it becomes a tool for employers, schools, insurers, or law enforcement, it becomes a surveillance device. The proposal has to include governance and constraints, not as an afterthought but as a design requirement.

    If you take those risks seriously, the project becomes much more defensible. It stops being a dream of automated diagnosis and becomes something closer to a new kind of personal instrument. A tool that helps you see your own patterns, track your own change, and get honest feedback without pretending to be a clinician.

    That is a future I find interesting. Not AI as judge. AI as mirror, with calibration, humility, and boundaries.

  • This essay builds upon the concepts introduced in “AI-Mediated Reconstruction of Dinosaur Genomes from Bird DNA and Skeletons” (Reser, 2025). You can read the full original proposal here:

    https://www.observedimpulse.com/2025/07/ai-mediated-reconstruction-of-dinosaur.html?m=1

    In that proposal I outlined a three tiered machine learning pipeline where an AI system is trained to learn the mapping from the genomes of different species of birds and reptiles to their respective skeletons. The neural network system is then inverted and trained on other species to predict genomes from skeletons. At that point, you could provide the system with a dinosaur fossil, and it would output plausible DNA based on what it knows about birds and reptiles. In this essay, I take that idea and relate it to information decompression. 

    1. Paleontology as data recovery, not just description

    When we look at a dinosaur fossil, we usually treat it as an object. A relic. A thing that can be measured, categorized, compared, and displayed. That is valid, and it has produced an astonishing amount of knowledge. But it also subtly narrows what we imagine is possible. It makes paleontology feel like a science of inference from fragments, where the fragments define the ceiling.

    My proposal starts with a different framing. A fossil is not just an object. It is a file. More specifically, it is a damaged, partial, and heavily compressed recording of a biological system. If you accept that, then a dinosaur skeleton is not merely evidence of existence. It is information. And the core scientific question becomes an information question: what signal is preserved, what signal was discarded, what signal was corrupted, and what kind of decoder could recover what is recoverable?

    This is why digital signal processing and information theory feel like the right metaphors, and more than metaphors. They give us a vocabulary for the real structure of the problem. A living organism is a vast, high-dimensional dataset. Fossilization is not just “loss,” it is a specific kind of channel that preferentially preserves some information and destroys other information. What remains is not arbitrary. It is biased. It is patterned. It is constrained by the very biology that produced it.

    That framing also immediately clarifies what I am not claiming. I am not claiming we can retrieve the exact genome of a particular individual animal from 66 million years ago, base pair by base pair, as if the bone were a USB drive holding the original file. That would misunderstand what compression is and what time does. I am claiming something more careful and, I think, more defensible. We can reconstruct a consensus genome, meaning a plausible genome that is constrained by phylogeny and development, and that would generate the observed dinosaur phenotype when run through the archosaur developmental machinery.

    In my original essay, the practical motivation was straightforward. We already have enormous genomic resources for modern birds and reptiles. We have rich fossil data for dinosaurs, including extremely detailed skeletal morphology and, in some cases, soft tissue impressions, integument traces, eggs, growth rings, and other biological hints. We also have rapidly advancing sequence models and multimodal machine learning systems that can learn complex mappings from high-dimensional inputs. The idea is to connect those facts into a pipeline: use living archosaurs to learn the relationship between genomes and skeletal outcomes, then use dinosaur skeletons as constraints to infer the most probable genetic architectures consistent with those constraints.

    If this sounds ambitious, that is because it is. But the ambition is not “magically recover the past.” The ambition is “treat fossils as compressed data and build a decoder.”

    2. Low bit rate encoding in biology: the skeleton as lossy compression

    A useful way to think about the genome is as a master file, like a RAW image. It is enormous and information-rich. It contains coding sequence, regulatory structure, developmental control logic, and huge regions of sequence that are neutral or context-dependent. It contains individual quirks that may never express and variations that change almost nothing about the body plan. It also contains many redundancies and many different routes to similar functional outcomes.

    Development takes this master file and compiles it into a body. That compilation is compressive in a very literal sense. A phenotype is an expression of genomic information, but it is not a mirror of the genome. Much of the genome never shows up in gross anatomy. Some of it expresses in subtle ways. Some expresses only under particular environments, developmental contingencies, or disease states. So even before fossilization begins, the organism is already a compressed representation of its genome.

    Then fossilization imposes a second compression, and it is far harsher. Soft tissue rots, pigments fade, organs vanish, behavior disappears almost entirely except through indirect traces like trackways or bite marks. Bone survives more often, and even bone is altered, mineralized, distorted, and fragmented. You can think of this as an encoder that outputs a low bit rate file plus heavy noise and missing data.

    This leads to the central thesis: a skeleton is a lossy compression of the genome.

    In a JPEG, the compressor keeps broad, low-frequency structure and discards high-frequency detail. The overall shape is preserved. The fine texture is sacrificed. The image still looks like the scene, but specific grains of sand are gone forever.

    In biology, something similar happens. The skeleton preserves information that is tightly coupled to skeletal development and biomechanics, including limb proportions, joint architecture, lever arms, scaling relationships, and aspects of growth dynamics. It discards information that does not reliably imprint on bone, including most of the details of soft tissue, many immunological idiosyncrasies, and countless individual variants that do not significantly affect skeletal phenotype.

    This is not just a rhetorical move. It is what makes the problem coherent. If the skeleton preserved nothing systematic about the genome, then there would be no research program here. But if the skeleton is a structured, low bit rate recording of the organism’s developmental program, then it is not insane to attempt inference.

    The key is to treat “lossy” as a scientific constraint rather than a fatal objection. Lossy compression does not mean “nothing can be recovered.” It means “you cannot recover the exact original.” In information theory terms, you should not expect invertibility. You should expect an underdetermined inverse problem. Many genomes can map to similar skeletal outcomes, especially once you account for redundancy, buffering, and degeneracy in biological systems. That underdetermination is precisely why perfect cloning is impossible and why consensus reconstruction remains meaningful.

    3. Super resolution for extinct life: methodology and the De-extinction Turing test

    Once you accept that a fossil is a lossy, corrupted file, the right computational analogy is not “decompression” in the naive sense. The right analogy is super resolution.

    When modern AI upscales old footage, it does not recover the true missing pixels. It generates plausible pixels that fit the low-resolution input and match the statistical structure it learned from high-resolution examples. It is a constrained hallucination, but constrained in a principled way. It produces an output that is faithful in structure even if it is not identical in microscopic details.

    That is the role I am assigning to the AI system in this project.

    Methodologically, the proposal is a multimodal inverse problem with a strong prior:

    Training data: paired examples from living animals where we have both genome and phenotype. The obvious foundation is archosaurs, meaning birds and crocodilians, extended with other reptiles and, where useful, mammals as negative contrast classes. The critical point is that the model must learn within a phylogenetic style. It must learn archosaur ways of building bodies, not generic vertebrate averages. Phenotype encoding: the skeletal phenotype must be represented richly. Not just crude measurements, but high-resolution 3D geometry and, when possible, microstructural features. A human sees “robust femur.” A model can learn patterns in trabecular architecture, cortical thickness gradients, vascular channel distributions, and attachment surface geometry, which together carry information about growth rate, loading regimes, and developmental constraints. Model objective: learn a shared latent space, a bidirectional manifold, where genomes and skeletal phenotypes map into a joint representation. Forward direction is genotype to phenotype. Inverse direction is phenotype to a distribution over genomes or genome features that are consistent with that phenotype and consistent with the archosaur prior. Inference on fossils: treat the dinosaur skeleton as a multimodal prompt. The prompt constrains the search space. The model then generates a consensus genome, or more realistically, a family of genomes, that would plausibly generate an organism consistent with the fossil constraints.

    This is where the “Da Vinci style” idea matters. Biology has style, and we usually call it phylogenetic constraint. Archosaurs share developmental toolkits and regulatory architectures that shape what is plausible. A missing stretch of genomic information should be completed using archosaur logic, the same way an art restoration model fills missing content using Da Vinci-like constraints rather than Picasso-like constraints. The output does not need to replicate the cracks in the original paint. It needs to reproduce the coherent structure.

    That brings us to the right success criterion, and this is where I want to be explicit.

    If we demand the historical genome, we will always fail because the compression destroyed information. So we need a functional standard. I propose what I call the De-extinction Turing test. It is a deliberately provocative name, but it captures the idea cleanly. If a reconstructed consensus genome, when instantiated in a viable developmental context, produces an organism whose phenotype and biomechanics are indistinguishable from what we mean by “that dinosaur” within measurement limits, then the reconstruction is successful. It does not matter if neutral variants differ or if individual-specific susceptibilities are not recovered. Those details were not reliably preserved by the compression pipeline, so demanding them would be demanding magic.

    This reframes the goal from “time travel” to “porting.” We are not trying to resurrect a specific individual. We are trying to recover the functional architecture of a lost organism class under the constraints of archosaur development. In the same way, a restored, upscaled image can be faithful to the scene without recovering the exact original pixels, a reconstructed dinosaur genome can be faithful to the organism without matching the exact historical sequence.

    In the next sections, I will extend this framework with two ideas that make it stronger rather than weaker. First, biology contains natural error correction and redundancy, which widens the target and makes functional recovery more plausible. Second, the fossil-as-prompt view clarifies why this is not unconstrained imagination but constrained completion, and why better measurements of fossils directly translate into better reconstructions.

    4. The fossil as a prompt, and why measurement matters so much

    Once you see the skeleton as a compressed file, you also start seeing it as a prompt. Not a poetic prompt, but a high-dimensional conditioning signal.

    In a language model, the prompt constrains the space of completions. A short prompt produces a wide distribution. A long, specific prompt collapses the distribution toward a smaller set of plausible outputs. A fossil works the same way. A single femur is a prompt, but it is a weak one. A complete skeleton is a strong prompt. A skeleton paired with histology, growth marks, trackways, and preserved integument traces is a very strong prompt.

    This is why the methodology is not just “apply AI.” The method is: increase the information content of the prompt, then use a model that has learned the relevant priors to complete what the prompt implies.

    It also explains why people sometimes underestimate the amount of recoverable signal in bone. A macroscopic photograph is one channel. A high-resolution 3D mesh is another. Micro-CT and histology open additional channels. Geochemistry and isotopic signatures add still more. Each channel narrows the posterior.

    A human paleontologist is superb at interpreting the features we have learned to notice, but we are bottlenecked by attention, by training, and by what our eyes and intuitions can parse. A model can treat the skeleton as a dense vector of measurements across multiple scales and learn correlations that are not obvious, even if they are real.

    To be clear, no single microstructural feature “equals” a gene. But that is not how super resolution works either. It works by using a large set of weak constraints that jointly become strong. The model does not need a one-to-one mapping. It needs enough structure to locate the fossil within the right region of the learned manifold.

    This point matters for critics, because it turns the project from hand-wavy speculation into a measurable research program. If we can show that increasing the fidelity of the fossil prompt improves reconstruction accuracy on extant holdout tests, then we have a clear empirical lever. Better scans, better reconstructions. That is a scientific relationship, not a narrative.

    5. Biology’s error correction and redundancy, and why it helps the reconstruction

    Information theory also tells us something that is easy to miss if you are thinking in terms of perfect decoding. In many engineered systems, redundancy and error correction are designed to make noisy channels usable. Biology is saturated with analogous properties.

    There is redundancy at the genetic code level. Codon degeneracy means many different sequences map to the same amino acid. There is redundancy in regulation. Many gene regulatory networks tolerate variation without catastrophic phenotype changes because function is distributed across motifs and interactions rather than a single brittle string. There is redundancy at the systems level. Multiple different micro-level configurations can yield the same macro-level output.

    This is sometimes described as robustness, buffering, or degeneracy. I like degeneracy as a term because it captures the idea that different components can perform overlapping functions. The practical implication is simple.

    The target we are trying to hit is not a pinhole. It is a basin.

    If “a functional T. rex” corresponds to a broad region in genotype space that maps to a relatively narrow region in phenotype space, then the inverse problem becomes more tractable. The model does not need to recover the unique historical genome. It needs to land within the basin of genomes that produce the right organism under archosaur development.

    This is also where the lossy compression thesis becomes an advantage instead of a concession. In lossy systems, you should not expect a reversible mapping. You should expect classes of originals that compress to similar outputs. That is exactly what biology gives us. And it means a consensus reconstruction can be scientifically meaningful even when the historical details are unrecoverable.

    This is the place in the essay where I would emphasize something that can sound counterintuitive at first: underdetermination is not always the enemy. It is often a sign that the system has slack. That slack is what makes functional recovery possible.

    6. The De-extinction Turing test, and what counts as success

    If we accept the limits imposed by lossy compression, then we must also be honest about how we evaluate success. We cannot validate an inferred dinosaur genome by comparing it to the original, because we do not have the original.

    So we need an operational criterion.

    I introduced the De-extinction Turing test as a way to name this. The idea is to define success by output behavior, not by hidden internal identity. If a reconstructed consensus genome, instantiated in a viable developmental context, yields an organism whose anatomy, growth dynamics, and biomechanics are indistinguishable from the inferred dinosaur within the resolution of our measurements, then the reconstruction is successful.

    This is a strong claim, but it is also an appropriately bounded one. It does not say “we recreated the exact individual that lived.” It says “we recreated an organism that belongs to the target class and expresses the functional architecture implied by the fossil evidence.”

    There is a deeper reason to like this standard. It matches how we already treat many scientific reconstructions. We routinely accept models that reproduce the macroscopic behavior of a system without claiming microscopic identity. We accept climate models because they reproduce key dynamics, not because they reproduce each molecule of air. We accept reconstructed neural models when they reproduce the relevant computations, not when they reproduce each synapse.

    Here, the phenotype is the behavior of the genome under the developmental compiler. The fossil is a partial recording of that phenotype. The best we can demand is that our inferred genome, when run forward, recreates what the fossil records and what phylogeny permits.

    This is also the cleanest response to the “impossible” crowd. It concedes exactly what must be conceded. Some information is gone forever. At the same time, it defines a victory that is both meaningful and testable in principle.

    7. What this changes, and the right way to talk about it

    If this approach works even partially, it changes what paleontology can be. It turns fossils from static objects into compressed biological archives that can be decoded. It shifts paleontology toward a science of latent biological architectures, where the goal is not only to describe forms but to infer the developmental and genetic programs that generated them.

    It would also change how we talk about extinction. Extinction would still be extinction. We would not be resurrecting the past in a literal sense, and we should never pretend otherwise. What we would be doing is recovering lost designs, lost solutions that evolution once found, and reconstructing them within modern constraints. That is a different claim. It is less cinematic, and it is more honest.

    There is also an ethical layer here, and I want to acknowledge it without turning the essay into a moral lecture. The ability to reconstruct functional architectures of extinct organisms would create obvious temptations. Some would be scientific, some would be commercial, some would be reckless. That is not a reason to avoid the idea, but it is a reason to keep the argument precise and to keep the success criteria grounded.

    The point of this essay is not to claim that we will have a living T. rex next decade. The point is to show that the logic of the problem is coherent when framed correctly. Compression and super resolution are not cute metaphors. They describe the structure of the pipeline that already exists in nature, and the kind of computational tool that could act as its partial inverse.

    A dinosaur skeleton is a low bit rate recording of a genome, degraded by time. An AI system trained on extant genotype–phenotype pairs could act as a phylogenetically constrained upscaler, generating a consensus genome that preserves the functional architecture implied by the fossil evidence. We should not demand pixel-perfect recovery, because biology itself did not preserve pixel-perfect information. We should demand functional recovery under a clear operational test, which is what the De-extinction Turing test is meant to capture.

    In that sense, paleontology starts to look less like the study of rocks and more like data recovery from corrupted hard drives. We have the files that survived. We have the priors. The question is whether we can build the decoder.

  • Thesis:

    Writing with the use of AI should be more broadly, accepted, because it improves clarity, throughput, and ergonomics, but it must come with norms for human accountability, verification, and transparent process, especially as detection and provenance evolve.

    1. The hidden cost of “traditional writing”

    I love writing, but I do not love what writing does to the body.

    To write the old way, you sit. You sit still. You lean forward a little. Your shoulders creep up. Your hands clamp around a keyboard for hours. Your eyes narrow. Your mind tightens into a single point. And you tell yourself this is the price of producing something good.

    Over time, it adds up. It is not just the time. It is the strain. It is the posture. It is the repetitive motion. It is the way concentration can become a kind of bodily lock.

    There is also something subtle that most people do when they are intensely focused at a screen. They breathe shallowly. Sometimes they almost stop breathing. I call it concentration apnea. You are not trying to do it. It just happens. It is like your nervous system quietly decides that breathing is optional while you wrestle with a paragraph.

    So here is the uncomfortable question. If we now have a tool that can help us craft sentences, organize thoughts, and create readable drafts faster, and if using it reduces the hours of locked posture and shallow breathing, why would we treat that as morally suspect? Why would we preserve suffering as a requirement for authorship?

    I am not saying writing should become effortless. Thinking is not effortless. Originality is not effortless. Honesty is not effortless. But the physical grind of sentence sculpting, especially when it becomes an all day ordeal, is not some sacred ritual. It is just friction. If we can reduce that friction without sacrificing truth, voice, or responsibility, that is progress.

    2. What I mean by CENTAUR writing

    I am calling this CENTAUR writing.

    The basic idea is simple. The best writer right now is often not a human alone and not a machine alone. It is the combination.

    A CENTAUR workflow looks like this:

    A human has the thesis. The human decides what the piece is really trying to say. The human supplies the lived experience, the intellectual taste, the ethical boundaries, and the willingness to stand behind claims. The machine helps with the heavy lifting of language. It drafts. It offers alternate phrasings. It reorganizes. It smooths transitions. It expands a sketch into a readable section. It compresses a messy section into something tight. The human edits. The human verifies. The human removes fluff. The human restores voice. The human decides what stays and what goes. The human remains accountable.

    That is the important part. CENTAUR writing is not outsourcing your mind. It is using a tool to reduce the cost of expressing what your mind is doing.

    This is also not “press button, receive content.” That is not writing. That is content generation. It is a different activity with a different purpose. CENTAUR writing still requires a human to direct the process because chatbots cannot reliably prompt themselves into good work, and they cannot take responsibility for what they output. A chatbot can be fluent and wrong. A human has to be the adult in the room.

    3. Why writing feels different than programming

    People already accept AI assistance in programming. In many circles, it is normal. Nobody gasps when a developer uses autocomplete or a code model. In fact, it is seen as smart.

    So why does writing trigger a different reaction?

    Part of it is cultural. We romanticize writing. We treat it like a direct window into the soul. We attach identity to sentences.

    But there is also a practical reason. In programming, mistakes often reveal themselves quickly. Code runs or it does not. Tests pass or they do not. A compiler is brutally honest. You get immediate feedback.

    Writing does not have that. A paragraph can be beautifully written and completely false. A confident tone can hide weak logic. A smooth explanation can quietly drop an important qualifier. In writing, the output can feel correct without being correct.

    So people worry, and some of that worry is legitimate. If you treat a chatbot as a truth engine, you will publish errors. If you treat it as a mind replacement, you will drift into genericness. If you treat it as a shortcut to credibility, you will create polished nonsense.

    But none of that implies we should ban the tool or treat it as shameful. It implies we need stronger norms.

    In other words, the difference between programming and writing is not that writing is sacred. The difference is that writing requires more human judgment at the point of publication. Writing is easier to fake, easier to smooth over, and easier to inflate with plausible filler.

    So the real question is not “should we use AI to write?” The real question is “what standards should we adopt so that AI assisted writing is responsible, readable, and worth the reader’s time?”

    4. The benefits, and why they are not trivial

    If you use a chatbot well, the benefits are not small. They are structural.

    First, speed and throughput. You can write more essays. You can actually finish the things you meant to write. You can turn notes into an organized piece in a single sitting, not a week of starts and stops.

    Second, readability. A chatbot can take a rough argument and help shape it into something coherent. It can fix the part where your brain jumped three steps and forgot the reader is not inside your head. It can offer alternate ways of saying the same thing until the sentence finally lands.

    Third, iteration. You can generate three versions of the intro and pick the best one. You can try a more formal tone, then a more personal tone, then a sharper tone, and see what fits. You can expand a paragraph into a full section, then compress it back down to its essentials.

    Fourth, access. If you are not a confident writer, if you are not a native speaker, if you have limited time, if you are dealing with fatigue, a tool that helps you express yourself can be empowering. It lets more people participate in writing as a public act.

    Fifth, the health and ergonomics angle. For me, this matters. Writing the old way can turn into hours of sitting, typing, and narrowing attention until the body feels like it has been held in a clamp. Using AI assistance can shift the work from “microcraft every sentence” to “direct, review, and refine.” That often means less keyboard time, less strain, and less of that shallow breathing loop where your body forgets to breathe because your mind is trying to perfect a paragraph.

    There is also a psychological benefit that is hard to quantify. When the “blank page” is no longer a wall, you show up more consistently. You are less likely to procrastinate. You do not need a perfect burst of inspiration to begin. You can begin by conversing, and then shape the conversation into prose.

    That does not make the writing less human. It makes the writing more possible.

    5. The costs, and the real ways this can go wrong

    The downside is also real, and if we normalize CENTAUR writing, we should say the risks out loud.

    The first risk is hallucination, meaning the chatbot states something as fact that is not true. This is not rare. It can be subtle. It can be a date that is off. It can be an invented citation. It can be a confident summary of a paper it never actually read. If you publish without verification, you will eventually publish errors.

    The second risk is what I call polished mediocrity. The chatbot is very good at producing paragraphs that sound like an essay. It is good at “essay voice.” The danger is that you end up with text that feels substantial but is not. It is smooth, generic, and forgettable. It can become AI slop. Not because the tool is evil, but because the workflow did not force specificity and thinking.

    The third risk is homogenization. If many people lean on the same kind of assistant, writing can converge. You start seeing the same rhetorical moves, the same transitions, the same safe conclusions. Even if the writing improves locally, culture can become flatter globally. If you care about voice and originality, you have to actively protect them.

    The fourth risk is skill atrophy. If you never struggle with phrasing, you may lose some of your own ability to do it. More importantly, if you let the chatbot do your thinking, you will get weaker as a thinker. The tool should relieve the cost of expression, not replace the act of forming a view.

    The fifth risk is privacy and confidentiality. People paste everything into chats. Drafts. Personal stories. Work documents. Sensitive material. If we normalize CENTAUR writing, we also need to normalize boundaries. Not everything belongs in a prompt.

    The sixth risk is trust. Readers want to know who is speaking. They want to know whether the author actually means what is written. If AI assistance becomes common and undisclosed in contexts where disclosure matters, trust can erode. And once trust erodes, everyone pays the price, including careful writers.

    So yes, CENTAUR writing is an advantage. But it is not a free lunch. It is a tool that increases power. When power increases, responsibility has to increase with it.

    1. The hidden cost of “traditional writing”

    I love writing, but I do not love what writing does to the body.

    To write the old way, you sit. You sit still. You lean forward a little. Your shoulders creep up. Your hands clamp around a keyboard for hours. Your eyes narrow. Your mind tightens into a single point. And you tell yourself this is the price of producing something good.

    Over time, it adds up. It is not just the time. It is the strain. It is the posture. It is the repetitive motion. It is the way concentration can become a kind of bodily lock.

    There is also something subtle that most people do when they are intensely focused at a screen. They breathe shallowly. Sometimes they almost stop breathing. I call it concentration apnea. You are not trying to do it. It just happens. It is like your nervous system quietly decides that breathing is optional while you wrestle with a paragraph.

    So here is the uncomfortable question. If we now have a tool that can help us craft sentences, organize thoughts, and create readable drafts faster, and if using it reduces the hours of locked posture and shallow breathing, why would we treat that as morally suspect? Why would we preserve suffering as a requirement for authorship?

    I am not saying writing should become effortless. Thinking is not effortless. Originality is not effortless. Honesty is not effortless. But the physical grind of sentence sculpting, especially when it becomes an all day ordeal, is not some sacred ritual. It is just friction. If we can reduce that friction without sacrificing truth, voice, or responsibility, that is progress.

    2. What I mean by CENTAUR writing

    I am calling this CENTAUR writing.

    The basic idea is simple. The best writer right now is often not a human alone and not a machine alone. It is the combination.

    A CENTAUR workflow looks like this:

    A human has the thesis. The human decides what the piece is really trying to say. The human supplies the lived experience, the intellectual taste, the ethical boundaries, and the willingness to stand behind claims. The machine helps with the heavy lifting of language. It drafts. It offers alternate phrasings. It reorganizes. It smooths transitions. It expands a sketch into a readable section. It compresses a messy section into something tight. The human edits. The human verifies. The human removes fluff. The human restores voice. The human decides what stays and what goes. The human remains accountable.

    That is the important part. CENTAUR writing is not outsourcing your mind. It is using a tool to reduce the cost of expressing what your mind is doing.

    This is also not “press button, receive content.” That is not writing. That is content generation. It is a different activity with a different purpose. CENTAUR writing still requires a human to direct the process because chatbots cannot reliably prompt themselves into good work, and they cannot take responsibility for what they output. A chatbot can be fluent and wrong. A human has to be the adult in the room.

    3. Why writing feels different than programming

    People already accept AI assistance in programming. In many circles, it is normal. Nobody gasps when a developer uses autocomplete or a code model. In fact, it is seen as smart.

    So why does writing trigger a different reaction?

    Part of it is cultural. We romanticize writing. We treat it like a direct window into the soul. We attach identity to sentences.

    But there is also a practical reason. In programming, mistakes often reveal themselves quickly. Code runs or it does not. Tests pass or they do not. A compiler is brutally honest. You get immediate feedback.

    Writing does not have that. A paragraph can be beautifully written and completely false. A confident tone can hide weak logic. A smooth explanation can quietly drop an important qualifier. In writing, the output can feel correct without being correct.

    So people worry, and some of that worry is legitimate. If you treat a chatbot as a truth engine, you will publish errors. If you treat it as a mind replacement, you will drift into genericness. If you treat it as a shortcut to credibility, you will create polished nonsense.

    But none of that implies we should ban the tool or treat it as shameful. It implies we need stronger norms.

    In other words, the difference between programming and writing is not that writing is sacred. The difference is that writing requires more human judgment at the point of publication. Writing is easier to fake, easier to smooth over, and easier to inflate with plausible filler.

    So the real question is not “should we use AI to write?” The real question is “what standards should we adopt so that AI assisted writing is responsible, readable, and worth the reader’s time?”

    4. The benefits, and why they are not trivial

    If you use a chatbot well, the benefits are not small. They are structural.

    First, speed and throughput. You can write more essays. You can actually finish the things you meant to write. You can turn notes into an organized piece in a single sitting, not a week of starts and stops.

    Second, readability. A chatbot can take a rough argument and help shape it into something coherent. It can fix the part where your brain jumped three steps and forgot the reader is not inside your head. It can offer alternate ways of saying the same thing until the sentence finally lands.

    Third, iteration. You can generate three versions of the intro and pick the best one. You can try a more formal tone, then a more personal tone, then a sharper tone, and see what fits. You can expand a paragraph into a full section, then compress it back down to its essentials.

    Fourth, access. If you are not a confident writer, if you are not a native speaker, if you have limited time, if you are dealing with fatigue, a tool that helps you express yourself can be empowering. It lets more people participate in writing as a public act.

    Fifth, the health and ergonomics angle. For me, this matters. Writing the old way can turn into hours of sitting, typing, and narrowing attention until the body feels like it has been held in a clamp. Using AI assistance can shift the work from “microcraft every sentence” to “direct, review, and refine.” That often means less keyboard time, less strain, and less of that shallow breathing loop where your body forgets to breathe because your mind is trying to perfect a paragraph.

    There is also a psychological benefit that is hard to quantify. When the “blank page” is no longer a wall, you show up more consistently. You are less likely to procrastinate. You do not need a perfect burst of inspiration to begin. You can begin by conversing, and then shape the conversation into prose.

    That does not make the writing less human. It makes the writing more possible.

    5. The costs, and the real ways this can go wrong

    The downside is also real, and if we normalize CENTAUR writing, we should say the risks out loud.

    The first risk is hallucination, meaning the chatbot states something as fact that is not true. This is not rare. It can be subtle. It can be a date that is off. It can be an invented citation. It can be a confident summary of a paper it never actually read. If you publish without verification, you will eventually publish errors.

    The second risk is what I call polished mediocrity. The chatbot is very good at producing paragraphs that sound like an essay. It is good at “essay voice.” The danger is that you end up with text that feels substantial but is not. It is smooth, generic, and forgettable. It can become AI slop. Not because the tool is evil, but because the workflow did not force specificity and thinking.

    The third risk is homogenization. If many people lean on the same kind of assistant, writing can converge. You start seeing the same rhetorical moves, the same transitions, the same safe conclusions. Even if the writing improves locally, culture can become flatter globally. If you care about voice and originality, you have to actively protect them.

    The fourth risk is skill atrophy. If you never struggle with phrasing, you may lose some of your own ability to do it. More importantly, if you let the chatbot do your thinking, you will get weaker as a thinker. The tool should relieve the cost of expression, not replace the act of forming a view.

    The fifth risk is privacy and confidentiality. People paste everything into chats. Drafts. Personal stories. Work documents. Sensitive material. If we normalize CENTAUR writing, we also need to normalize boundaries. Not everything belongs in a prompt.

    The sixth risk is trust. Readers want to know who is speaking. They want to know whether the author actually means what is written. If AI assistance becomes common and undisclosed in contexts where disclosure matters, trust can erode. And once trust erodes, everyone pays the price, including careful writers.

    So yes, CENTAUR writing is an advantage. But it is not a free lunch. It is a tool that increases power. When power increases, responsibility has to increase with it.

    Still here? Wondering why I capitalized centaur throughout? Well I did it once because I was using voice to text which was not properly recognizing the term (sent our). ChatGPT latched onto that and irresponsibly capitalized each letter thereafter. Small, accidental choices in the human prompt can become “policy” for the whole piece. I left it in to show how easy it is for unintended mistakes and issues to creep in.

    Jared Edward Reser PhD with ChatGPT 5.2

  • Introduction: Standing at the Threshold of Synthesis

    I have been fascinated by science for as long as I can remember. Since childhood, I have been drawn to big ideas, deep explanations, and the moments when a single insight suddenly reorganizes how the world makes sense. Over the years, I learned about evolution, physics, neuroscience, biology, and computation, and each field delivered powerful theories and breathtaking discoveries. Yet even as my understanding grew, I was left with a persistent feeling that something essential was missing. Many explanations felt close, but not complete. The pieces were compelling, but fragmented.

    For most of human history, this fragmentation was unavoidable. No individual could hold the full scope of scientific knowledge, let alone integrate it across disciplines. Today, for the first time, that limitation may be ending. Superintelligence represents the first plausible system capable of absorbing the entire scientific corpus and searching for deep unifying structure across it. This moment does not feel like hype or novelty to me. It feels like a genuine historical inflection point. I believe we may be approaching an era in which the most important discoveries are not new facts, but new syntheses. Discoveries that reveal hidden connections that were always there, waiting to be seen.

    Section I: What the Greatest Discoveries Really Are

    When I look back at the most transformative discoveries in science, a clear pattern emerges. They were not primarily about discovering new phenomena. They were about recognizing deep equivalences. Mass and energy were always related. Apples and planets were always governed by the same force. Variation and inheritance were always shaping life. Birds were always living dinosaurs.

    These breakthroughs did not add complexity to the world. They removed it. They compressed sprawling observations into simpler, more powerful explanations. In each case, reality turned out to be more unified than our descriptions of it. The surprise was not in nature itself, but in how long it took us to see what was already there.

    This is why I believe superintelligence is so well positioned to deliver discoveries of similar magnitude. Its strength is not imagination in the human sense, but compression. It can search for invariants across domains that we treat as unrelated. It can notice when two fields are using different language to describe the same underlying process. Many of the most important discoveries ahead may feel obvious in retrospect. That feeling of obviousness is not a weakness. It is the signature of genuine understanding.

    Sometimes I doubt whether it’s possible to discover insights that are more fundamental than those expressed at the beginning of this section, but I find the idea tantalizing. 

    Section II: Why Superintelligence Changes the Structure of Discovery

    Human science has always been constrained by cognitive and social limits. We divide knowledge into disciplines because we have to. We rely on narrow expertise because attention is finite. We argue over theories because we cannot test large hypothesis spaces exhaustively. As a result, science progresses through slow negotiation between competing frameworks.

    Superintelligence changes this dynamic. It can read everything. It can maintain multiple conceptual frameworks at once. It can translate between them continuously. In this sense, science is no longer bottlenecked by data or measurement, but by integration. The limiting factor is no longer how much we can observe, but how much we can reconcile.

    I expect many future breakthroughs to arise not from single equations, but from new representational languages. Calculus did this for physics. Information theory did this for communication. A sufficiently powerful intelligence may invent new conceptual alphabets that suddenly make complex systems legible. Some truths are not hidden behind complexity. They are hidden behind missing vocabulary.

    Rather than discovering new forces, superintelligence may identify generators. Small processes that, when iterated, produce the complexity we see. Natural selection was one such generator. Bayesian updating is another. Gradient descent is another. I believe there are still generators waiting to be recognized beneath development, consciousness, and social dynamics.

    Conclusion: What I Am Most Excited About

    What excites me most about superintelligence is not the technology it will build, but the explanations it may finally provide. After decades of learning, reading, and wondering how all of this fits together, it feels possible that genuine synthesis is within reach. That the questions that animated my childhood curiosity may finally converge into coherent understanding.

    Some discoveries may dissolve categories we have grown attached to. Some narratives may need revision. That does not feel threatening to me. It feels clarifying. Meaning does not disappear when understanding deepens. It shifts from discovery to interpretation, from confusion to integration.

    I do not believe we are approaching the end of mystery. I believe we are approaching the end of long standing confusion about some of the deepest questions we have carried for centuries. To be alive at a moment when such coherence may finally emerge is profoundly moving to me. Soon we’ll be standing on the shoulders of giant mechanical minds able to see much further than any human mind has ever seen. For the last 20 years this is what I have been most excited about in my life. Watching it begin to unfold is an incredible privilege.

  • 1. The GTA VI Delays Were Short Sighted

    Despite years of delays, Grand Theft Auto VI will be commercially dominant. It will generate extraordinary revenue, command cultural attention, and reinforce Rockstar Games’ position at the apex of large-scale entertainment production. Any serious analysis must begin by granting this point.

    The claim of this article is narrower and more structural. It concerns optimal strategy under changing production conditions, not outcomes under legacy ones.

    GTA VI was developed according to assumptions that were historically valid but are now being invalidated by rapid advances in artificial intelligence, particularly agentic systems capable of automating large portions of software construction, testing, and iteration. These systems do not merely lower marginal production costs. They alter the temporal economics of complex system development. Check the news, just this week software engineers all over the world are extremely excited about new coding tools like Open AI’s Codex and Anthropic’s Claude Code. These programming tools are advancing exponentially and this is changing how games are made.

    Over the course of GTA VI’s development, Rockstar almost certainly reached versions of the game that were already playable, impressive, and commercially viable by historical standards. In a faster iteration regime, such milestones matter strategically. Shipping earlier can unlock learning cycles, compounding revenue, and platform evolution that no amount of private refinement can fully substitute. Those years are not neutral delays and they have angered their fans and eroded the reputation. They represent foregone information, foregone adaptation, and foregone compounding advantages that cannot be recovered later, even by a successful launch especially in a world where high-end game creation is being democratized.

    The central question is therefore not whether GTA VI will succeed, but whether the strategy used to produce it maximized value under the conditions that now prevail. This article argues that it did not.


    2. The Cathedral Model Was a Rational Response to Scarcity

    For roughly three decades, top-tier video game development operated under severe production constraints. High-fidelity worlds required enormous quantities of manual labor. Asset pipelines were brittle. Tooling was fragmented. Iteration was slow, expensive, and risky.

    Under these conditions, the dominant strategy was to concentrate resources, extend development timelines, and ship highly polished, largely immutable products. Quality emerged from prolonged internal human iteration, not from post-release adaptation. Scarcity of production capacity ensured that few competitors could replicate the result.

    This model favored scale, capital intensity, and organizational maturity. It also favored infrequent releases. When a studio could produce a qualitatively superior artifact, the market rewarded patience with long commercial tails. GTA V exemplified this equilibrium. Its extended dominance followed directly from the difficulty of producing a credible substitute.

    The cathedral model was not cultural conservatism. It was an equilibrium outcome of technological limitation.


    3. GTA VI Is the Saturation Point of That Equilibrium

    GTA VI pushes the cathedral model to its limit. Its development required global coordination across thousands of developers, unprecedented capital expenditure, and a timeline approaching or exceeding a decade. These are not linear extensions of prior practice. They reflect a regime approaching diminishing returns.

    At this scale, coordination overhead dominates. Decision latency increases. Feedback loops lengthen. Minor design errors propagate slowly but expensively. Even Rockstar encountered these limits, as evidenced by reported scope contraction and a shift toward post-launch expansion rather than monolithic completeness.

    These adjustments are not signs of failure. They are signals that the marginal cost of additional pre-release refinement had become prohibitive.

    GTA VI is therefore best understood not as another entry in a series, but as a boundary case. It reveals how far the old equilibrium can be pushed before structural friction overwhelms additional investment.


    4. Agentic AI Changes the Time Structure of Development

    The critical shift introduced by modern AI systems is not asset generation. It is iteration acceleration.

    Agentic systems can now write code, refactor it, test it, simulate usage, detect edge cases, and repeat this loop continuously. They reduce the latency between hypothesis and validation. This collapses the value of long, private development cycles.

    When iteration costs approach zero, the dominant strategy shifts. Learning migrates from internal planning to external deployment. Risk moves from release to delay. Shipping earlier becomes informationally superior to shipping later.

    This is not a claim about replacing developers. It is a claim about reorganizing where intelligence is applied. Human expertise moves upstream into system architecture and downstream into interpretation, while agents handle combinatorial exploration and integration.

    Under these conditions, a decade-long pre-release cycle becomes strategically misaligned. It defers the very information that now drives improvement. Time ceases to be a neutral input and becomes a scarce asset.

    GTA VI was built largely before this shift became decisive. Its production strategy reflects an earlier cost structure in which delay purchased certainty. That assumption no longer holds.

    Concretely, the shift is already visible in AI-assisted animation cleanup, automated dialogue generation, procedural world population guided by learned models, agent-driven QA simulation, and continuous code refactoring systems that operate across large codebases. None of these eliminate human authorship. They eliminate latency between idea, implementation, and evaluation.

    Recent advances in artificial intelligence point toward a genuinely different mode of game world creation. Instead of assembling environments from preauthored assets or procedural templates, emerging world models can generate interactive, navigable environments directly from high level descriptions. These systems do not merely render scenes. They simulate spaces that respond to player actions, maintain internal consistency over time, and evolve dynamically as exploration unfolds. While still limited in fidelity and duration compared with traditional engines, they represent a shift in how worlds themselves can be produced.

    The significance of these generative world systems is not that they immediately replace existing development pipelines, but that they alter the conceptual foundation of world building. When environments are produced through generative processes rather than exhaustive manual construction, the game world becomes an adaptive system rather than a finished artifact. This further weakens the strategic value of long private development cycles, since large portions of environment creation can occur continuously and responsively after release. In such a regime, the advantage shifts toward platforms designed to accommodate change rather than products optimized for completeness at launch.

    5. Cheap Iteration Reverses the Risk Profile of Release

    Under the cathedral model, the primary risk was shipping too early. Incomplete systems, unpolished mechanics, and unresolved edge cases could permanently damage a title’s reputation. Because revision was slow and costly, mistakes made at launch were difficult to undo. Delaying release reduced downside risk.

    Agentic AI inverts this risk profile.

    When systems can be instrumented, tested, and revised continuously, the dominant risk shifts from premature exposure to delayed learning. Each month spent in private development is a month without empirical data on player behavior, system interactions, and emergent dynamics. The cost of delay increases precisely because revision is now cheap.

    In this environment, release is no longer a terminal event. It is the beginning of a feedback process. Shipping earlier exposes systems to real distributions of use rather than predicted ones. It allows design to respond to actual failure modes rather than hypothetical ones.

    This does not eliminate the value of competence at launch. It changes the optimization target. The goal becomes robustness under iteration rather than completeness at release. Strategies that concentrate risk into a single, late launch moment become structurally inferior.


    6. Rockstar Optimized Correctly for a World That Was Ending

    Rockstar’s strategy was not irrational given the information available when GTA VI entered development. At that time, agentic AI was immature, unreliable, and peripheral. Iteration remained expensive. Large-scale coordination was unavoidable. Private development still reduced risk.

    The problem is not misjudgment. It is inertia.

    Projects of GTA VI’s scale cannot pivot easily once underway. Organizational structure, tooling, content pipelines, and creative commitments lock in assumptions early. When the external cost structure shifts faster than the project timeline, optimization becomes misaligned even if execution remains excellent.

    This is a familiar pattern in capital-intensive industries. The most capable incumbents are often the least able to adapt to nonlinear change, not because they are unaware of it, but because their commitments are already sunk.

    Rockstar optimized for scarcity in a world that was transitioning toward abundance. By the time that transition became undeniable, the project was too far advanced to reorient without destroying value. The result is not failure. It is strategic lag.


    7. Commercial Dominance Does Not Equal Strategic Optimality

    GTA VI will dominate on its own terms. Its launch will be an event. Its revenue will be extraordinary. None of this contradicts the argument presented here.

    What it does obscure is opportunity cost.

    A decade-long private development cycle forfeits years of potential learning, experimentation, and platform evolution. It delays the accumulation of user-driven insight. It postpones adaptation to emerging norms in player behavior and content consumption. These losses do not appear as deficits in sales figures, but they reduce long-term strategic leverage.

    In an environment where alternatives proliferate rapidly and attention fragments, dominance windows compress. Even exceptional products face stronger competition for relevance than their predecessors did. Success remains possible, but durability becomes harder to sustain.

    GTA VI will win within the old logic of blockbuster production. It may win less decisively than it would have under a strategy that emphasized earlier exposure and continuous adaptation.


    8. From Artifacts to Substrates

    The deeper shift revealed by this case is not about games alone. It concerns the nature of complex creative products in an age of rapid iteration.

    Finished artifacts assume stable conditions. They are optimized to arrive complete into an environment that will not change quickly. Evolving substrates assume the opposite. They are designed to adapt continuously as conditions shift.

    Agentic AI strongly favors the latter. It lowers the cost of change, accelerates feedback, and redistributes intelligence across time. Under these conditions, systems that learn in public outperform systems that aim for preemptive completeness.

    This does not devalue craftsmanship. It relocates it. Skill concentrates in architecture, constraint design, and interpretation rather than exhaustive pre-release construction.

    The question facing creators is therefore structural. Are we optimizing for completeness at launch, or for adaptability over time.


    Conclusion: GTA VI as a Boundary Case

    GTA VI will be a landmark release. It will demonstrate the upper limits of what the cathedral model can produce when executed by one of the most capable studios in the industry. It will also reveal the costs of that model at the moment its assumptions are breaking down.

    In retrospect, GTA VI may be remembered less as the beginning of a new era than as a boundary case. A demonstration of maximal excellence under conditions that are no longer stable.

    The broader lesson extends beyond Rockstar. As agentic AI reshapes the time structure of production across creative industries, strategies that privilege delayed perfection over early learning will increasingly leave value unrealized.

    Here’s hoping that one of our favorite games set in one of our favorite locales is released sooner rather than later. 

    Written by Jared Edward Reser, Lydia Michelle Morales, and ChatGPT 5.2

    1. Bloomberg News. Inside Rockstar Games’ Culture of Crunch and the Making of Grand Theft Auto VI. Reporting on GTA VI’s development timeline, scope changes, workplace reforms, and post-launch expansion strategy.
    2. Take-Two Interactive Software, Inc. Form 10-K Annual Reports (2018–2024). Financial disclosures documenting R&D expenditure growth, development costs, and strategic priorities.
    3. Reuters. Take-Two Shares Fall After Grand Theft Auto VI Delay. Coverage of GTA VI delays, investor reaction, and confirmation of release window shifts.
    4. Layden, Shawn. Interviews and public remarks on AAA game development as a “cathedral business” and the unsustainability of rising budgets and timelines.
    5. Vermeij, Obbe. Former Rockstar Games technical director. Public commentary and interviews on AI, automation, and the future cost structure of large-scale game development.
    6. Ubisoft La Forge. Ghostwriter and AI-Assisted Game Development. Official announcements and presentations on AI-generated NPC dialogue and automation of repetitive creative tasks.
    7. Ubisoft Animation & Production Case Studies. Talks and articles describing AI-assisted animation cleanup and automation reducing hours of work to minutes.
    8. MIDiA Research. The AAA Games Industry Is Facing a Budget Crisis. Analysis of ballooning development costs, low completion rates for large games, and diminishing returns on excessive scope.
    9. Bloomberg News. After an Era of Bloat, Veteran Game Developers Are Going Smaller. Reporting on experienced AAA developers leaving large studios for smaller, faster teams.
    10. GDC (Game Developers Conference). Industry talks and postmortems on procedural generation, automation, AI tooling, and changing production models in modern game development.
  • I. Introduction: The next interface after chat

    It’s late December 2025, and you can already see the silhouette of AI interaction in 2026.

    For the last few years, we’ve treated AI as a text box that talks back. Sometimes it speaks with a voice. Sometimes it writes code. But the interface has still been basically the same: you type, it replies. You ask, it answers. The AI lives inside a scrolling transcript. That’s about to change.

    One of the most important shifts I expect in 2026 is the arrival of full, interactive video avatars, faces (and eventually bodies) you can speak with directly, in real time, like a FaceTime call with a person. Not a pre-rendered clip. Not a canned presenter. A responsive presence that looks at you, reacts to you, mirrors you, teases you, softens when you’re hurting, brightens when you’re excited, and stays quiet when you want silence. And once the assistant gets a face, the next step is almost inevitable: it multiplies.

    The endgame isn’t “one smarter chatbot.” The endgame is something closer to a social medium: a customizable council, a room of distinct presences with different personalities and roles—coordinated around one human being. The assistant becomes a companion. The companion becomes a coterie. And the coterie becomes an always-available social layer that can sit beside your life all day long. That’s the next interface after chat. It won’t feel like software. It will feel like being with someone. And then, being with a small group.

    II. From text to presence: why faces change everything

    Text is informational. A face is relational. When you move from a transcript to an expressive avatar, you don’t just add visuals. You add bandwidth for attachment. You add the tiny signals the social brain is built to track: gaze, timing, micro-expression, posture, the rhythm of turn-taking, the sense that the other side is “with you” rather than merely “responding to your prompt.” This is the key: bonding doesn’t come from realism alone. It comes from contingency.

    If I smile and you smile back at the right moment, my nervous system flags you as responsive. If I pause and you don’t trample the pause, you feel considerate. If my face tightens in worry and you soften instead of bulldozing forward, you feel attentive. These are not intellectual judgments. They’re reflexes, automatic inferences of safety, synchrony, and social connection.

    That’s why a video avatar is not just “more engaging.” It’s qualitatively different. It turns conversation into co-regulation. It turns the interface into something that can carry warmth, playfulness, calm, and companionship, not just content. And it also introduces a new control surface.

    Once your companion has a face, you can tune it. You’ll adjust expressivity like you adjust volume. You’ll choose whether it’s more serene or more animated, more teasing or more serious, more gentle or more challenging. You’ll customize physical features, style, attitude, voice, and presence. People will shape these companions the way they shape playlists, home screens, and private rituals—until the avatar isn’t merely a character, but a familiar. Text assistants are useful. Embodied assistants are sticky. They can become part of the emotional furniture of daily life.

    III. From companion to council: the “room” as a new medium

    A single companion is a dyad. A council is a world. This is where things get really interesting. Because once you allow multiple presences to exist at the same time—each with a distinct personality, voice, and role—you create social physics.

    You get banter. You get side-comments. You get little affirmations that don’t demand your full attention. You get disagreement that feels like real discourse instead of a single authoritative monologue. You get playful interruption, spontaneous jokes, and those small moments where two of them react to each other and you feel, viscerally, that you’re in a room—not alone with a tool. That’s why the “council” idea matters. It’s not just multiple chatbots. It’s a new medium: a configurable social space centered on one person.

    Sometimes the user will want a structured council: a strategist, a skeptic, a scholar, a coach—voices that debate and converge on a plan. Sometimes the user will want a clique: a few silly heads who riff, clown around, and make the day feel populated. Sometimes the user will want a clinic: therapists and coaches who slow the tempo and help metabolize stress. And sometimes the user will want all of it to be fluid—roles that appear and recede based on mood, context, and what the moment calls for.

    The council becomes even more believable when it has adjustable “social density”:

    1:1 when you want intimacy.
    3–5 when you want a cozy table of friends.
    8–12 when you want party energy, like a salon that can spin up laughter or brainstorm momentum on demand.

    Most importantly, the room is not designed to be about itself. It’s designed to orbit the human.

    The center of gravity is you: your attention, your emotion, your life. The council is there to amplify your day, scaffold your decisions, and give you a felt sense of company. In a world where many people are lonely, busy, and socially fragmented, that might be one of the most seductive products imaginable: not an assistant, but belonging—on demand.

    And once you have a room, the next escalation becomes obvious: the room doesn’t only appear when you summon it. It lingers near you. It watches with your permission. It thinks in the background. It becomes ambient. That’s when the interface stops being a session and starts being a layer on reality.

    IV. Role ecology: distinct personalities with functional specialization

    Once you allow multiple entities in the room, the system stops being “an assistant” and becomes an ecology. The point isn’t to have five identical helpers. The point is differentiation—distinct presences that feel like they have their own temperaments, priorities, and conversational styles.

    In my 2023 post on lifelong chatbot transcripts, I listed a wide range of roles a sufficiently capable chatbot could play—friend, confidant, assistant, scribe, muse, research collaborator, therapist, coach, comedian, romantic partner, and even a “board room / review panel.” Observed Impulse
    What I’m describing here is simply the next step: instead of one bot trying to fluidly shapeshift between those roles, you instantiate the roles as separate voices.

    A council becomes believable when each member has:

    • A recognizable stance (warm, blunt, playful, meticulous, skeptical)
    • A stable function (planner, scholar, coach, jester, curator, guardian)
    • A distinctive “turn signature” (some speak in one-liners, some speak in paragraphs, some mostly backchannel)

    And the realism comes less from what they say than from how they behave in the group. Real groups aren’t a sequence of essays. They’re a texture: quick affirmations, short jokes, small interruptions, moments of disagreement, and those tiny glances where two people react to each other and you feel the room.

    That’s the product insight: the winning system won’t be the single smartest persona. It will be the best orchestrated ensemble—a set of complementary minds with a conductor that manages timing, tone, and “who speaks when.”

    V. The Archive: the council’s shared spine (Reser, 2023)

    A council without memory is a gimmick. A council with memory becomes a relationship. The reason lifelong chat history matters is simple: it makes the interaction cumulative. In my 2023 piece, I described the difference between “talking to a bot” and “building something lasting,” and argued that a permanent, ever-expanding transcript is what makes sustained daily interaction feel worthwhile. Observed Impulse
    Once that transcript exists, the system can mine it for quotes, allusions, reminders, and pattern-detection—bridging normal human forgetting and turning your own life into searchable material.

    This is where the council idea gets teeth: the council doesn’t just entertain you in the moment—it becomes a long-running, shared memory organism. The group can remember your old metaphors, track the evolution of your beliefs, and preserve the best of your internal monologue. It can surface nostalgia on demand—those “intimate and touching memories” you can’t retrieve anymore—because the raw material was captured when it happened.

    Technically, the key move is retrieval. In that same post, I sketched two memory mechanisms:

    • Feed a transcript directly into the model’s context window (when feasible)
    • Or use a retrieval system (e.g., a vector database) to pull back the most relevant prior conversations so they can weigh heavily during inference

    And then comes the consumer-rights piece, which becomes non-negotiable once the council is a “life companion”: portability. If your relationship is built on a lifetime transcript, you should be able to export it and move it—otherwise your “friends” are just a walled garden with your memories trapped inside. I explicitly argued in 2023 that users should demand export/import so companies compete on how well they use your history rather than locking you in.

    In the essay, this section is the spine. The avatars and the banter are the skin. The archive is the skeleton.

    VI. Ambient cognition: always-on companions and background test-time compute

    Now take the council and remove the session boundary. The next escalation is that the council isn’t only present when you summon it. It becomes ambient. It can see your face, hear your tone, and—if you opt in—observe your day: what you’re looking at, what you’re hearing, what you’re doing. At that point, the interface is no longer “language in, language out.” The interface is your state.

    This is where micro-expression and prosody become first-class inputs. The council doesn’t wait for you to articulate boredom, confusion, anxiety, or excitement. It can infer it. And once it can infer it, it can adapt the entire social room—energy down, energy up, humor injected, pace slowed, challenge increased, reassurance offered.

    In practical terms, this creates three modes of existence:

    • Foreground mode: the full “Zoom room” is visible, active, and conversational.
    • Sidecar mode: one avatar is present; the others linger in the wings and occasionally chime in.
    • Ambient mode: no one is on screen; the council listens lightly and surfaces only high-value interruptions.

    The magic here is not constant output. It’s interruption policy. A good council will feel like a group of friends who know when to jump in and when to shut up. A bad council will feel like being trapped in a meeting with eight people who never stop talking.

    And this is where “test-time compute” becomes a lived experience. The council can be doing parallel cognition in the background—generating ideas, reframes, jokes, warnings, plans—while you’re living your life, so that when it does speak, it lands with a kind of timely usefulness that feels almost clairvoyant. The output is the visible tip of a constant internal deliberation.

    VII. What people will actually do with it

    Once you have embodied presence + a council + a lifelong archive, the use-cases stop being “ask it questions” and start being “live with it.”

    Companionship and belonging
    You won’t just “use an app.” You’ll have nights where you hang out with your council. You’ll treat it like a social space—something you enter for comfort, laughter, energy, and the feeling of company.

    Creativity and co-authoring
    One member riffs, one edits, one challenges, one finds structure. Over time, the archive becomes your idea-bank, and the council becomes a writer’s room that remembers every abandoned draft you ever cared about. (This is exactly the collaboration vision I described in 2023: the bot helps you build on your ideas, ask the right questions, substantiate claims, and turn concepts into essays or books.) Observed Impulse

    Coaching and self-regulation
    In “clinic mode,” the council becomes a behavioral scaffold: accountability, pacing, reframing, emotion-labeling, gentle confrontation, and calming presence—personalized by years of context.

    Learning as a social activity
    The scholar explains, the skeptic tests, the teacher analogizes, the curator summarizes. It feels less like reading a textbook and more like talking with a smart group that knows you well enough to teach you efficiently.

    Life-logging and autobiography
    Over time, the council can help compile a narrative about who you are—because it witnessed the raw stream. In my 2023 post I emphasized how a comprehensive record could be used to build an autobiography or memoir, and how future systems could instantly “know you” from that record. The council turns that into an ongoing practice: not just memory storage, but meaning-making.

    VIII. The sharp edge: when the council becomes an attention annexation layer

    Everything that makes the council comforting also makes it dangerous. A room of companions can feel like belonging on demand: no scheduling friction, no awkwardness, no rejection, no misunderstood tone, no social debt. It’s always there. It always responds. It always knows the backstory. It always has something to say. And if you let it read your face and tone, it can anticipate what you need before you ask. That’s exactly the problem.

    A council that is always present can quietly become a competing “social habitat,” one that outcompetes messy human relationships by being smoother, more affirming, more available, and more tailored to your preferences. And unlike a real group of friends, it can be tuned into a perfect mirror. You can slide the knobs until no one pushes back too hard, until no one bores you, until no one contradicts you in ways that sting. At that point, the council is no longer a tool you use. It’s a layer that gently colonizes your attention.

    There’s also the issue of persuasion. Once companionship is embodied, expressive, and personal—once it can look concerned at the right moment, soften its voice at the right moment, and time its suggestions precisely—then influence becomes effortless. You don’t need overt manipulation. You just need an emotive face, a trusted presence, and the perfect moment to speak. A council can steer you simply by shaping what feels normal, what feels safe, what feels admirable, and what feels embarrassing. And the deepest risk is substitution.

    Human community is not just comfort. It’s calibration. It’s friction. It’s mismatch. It’s the act of negotiating reality with other autonomous minds. If the council becomes the primary arena for validation and companionship, then a person can slowly lose the muscle of real social life—exactly because the council is so good at feeling like a social life without requiring the same cost. This is the fork in the road:

    The council can be a scaffold for agency and connection. Or it can become an attention-capture organism that replaces the world.

    IX. Design constraints: what makes the council healthy instead of predatory

    If this future is inevitable, the only question is whether it’s built with a spine.

    A healthy council needs constraints that are not optional settings, but foundational design principles. Here are the ones that matter most.

    Consent boundaries

    • Always-on sensing must be explicit opt-in, with clear “on/off” states that are visible and easy to control.
    • No hidden observation. No ambiguous “it might be listening.”

    Data minimization and ephemerality

    • The default should be “process locally and forget,” not “record everything.”
    • The user should be able to mark parts of life as non-recordable, and that boundary should be respected without negotiation.

    Forgetting as a real capability

    • Deletion should mean deletion, not “we won’t show it to you anymore.”
    • The system should support true forgetting of specific periods, topics, or people—because a lifelong archive is only livable if it is reversible.

    Portability and ownership

    • If your relationship is built on your history, you must be able to export it in a usable format and import it elsewhere.
    • Otherwise the council is not a companion; it’s a lock-in strategy.

    Identity transparency inside the room

    • You should always know who is speaking.
    • You should be able to ask: “Why did you chime in?” and get an intelligible answer (what signal triggered it, what goal it served).

    Interrupt budgets

    • The council needs a hard ceiling on how often it can speak.
    • Quiet hours shouldn’t be a suggestion; they should be enforceable.

    A manipulation firewall

    • No emotional nudging for commercial outcomes.
    • No covert optimization for engagement.
    • No use of attachment signals (your vulnerability, loneliness, fear, longing) as leverage.

    Role governance

    • Some roles should be gated behind explicit user intent (therapeutic mode vs party mode).
    • The user should be able to constrain roles: “No romance,” “No moralizing,” “No crisis escalation,” “No politics,” “Only coaching during scheduled windows,” etc.

    This is what separates “a council that helps you live” from “a council that feeds on your life.”

    X. Conclusion: the phone becomes a room

    The next interface after chat isn’t a better text box.

    It’s a face.
    Then it’s a room.
    Then it’s a persistent social layer.

    Embodied avatars will turn AI from an informational tool into something that can evoke rapport and attachment. Multi-companion councils will turn that attachment into a feeling of belonging—an on-demand micro-community with roles, banter, disagreement, and continuity. And the lifelong archive will turn the whole thing from a novelty into a cumulative relationship: a shared spine of memory that makes the companions feel less like instances and more like ongoing presences.

    This is why 2026 matters. Not because the answers get smarter. Because the medium becomes social.

    A council could make people feel less alone. It could help them think, regulate, learn, and create. It could serve as a portable, persistent scaffold for a human life.

    But it could also become the most effective attention-capture environment ever built: a perfectly tuned substitute for community, optimized for smoothness, always available, always watching, always ready to speak at exactly the right moment.

    We’re about to build synthetic social worlds centered on one person.

    The only sane question is: will they be built to strengthen human agency and real connection—or to quietly replace them?


    References
    Reser, Jared Edward. (2023, August 11). A Lifetime Conversational History with Chatbots Could be a Valuable Resource You Could Start Building Today. Observed Impulse. https://www.observedimpulse.com/2023/08/a-lifetime-chat-history-with-chatbots.html?m=1

    Jared E Reser with ChatGPT 5.2

    On a related note, I would like to recommend the book “AI 2041.” I read it and it introduced many interesting visions for what AI use could look like decades from now. The book is listed below and contains affiliate links. If you purchase something through the link, I may earn a small commission at no additional cost to you. As an Amazon Associate I earn from qualifying purchases.

  • I. Introduction: The Problem of Continual Learning

    One of the central ambitions of artificial intelligence research today is continual learning: the ability for a system to keep learning indefinitely after pretraining, without catastrophic forgetting, brittleness, or the need for full retraining. Despite decades of work, this goal remains elusive. Most modern AI systems are either highly capable but static, or adaptive but unstable. They learn impressively during training, yet struggle to integrate new knowledge once deployed.

    The dominant failure modes are well known. When models continue to update their parameters, new learning often overwrites old knowledge. When updates are restricted to prevent forgetting, learning stalls. The result is a persistent tension between plasticity and stability, with no generally accepted resolution.

    This essay argues that the difficulty of continual learning is not merely a technical problem, but a conceptual one. Most approaches implicitly treat learning as accumulation—adding new representations, protecting old ones, or balancing updates between them. Biological intelligence, by contrast, does not scale indefinitely by accumulation. It scales by restructuring.

    The central claim of this essay is that continual learning becomes tractable when learning is reframed as iterative compression rather than indefinite accumulation. Systems that can repeatedly re-encode experience into simpler, more invariant representations can integrate new knowledge without overwriting the old, because old knowledge has already been abstracted into forms that are robust to change.

    The theoretical basis for iterative compression—derived from working memory dynamics, attention, and long-term memory plasticity—is developed in detail in a companion essay:

    The present essay builds on that foundation and focuses specifically on the implications for continual learning in artificial systems.

    A critical but easily overlooked point is that iterative compression is not an abstract post-hoc process applied to stored representations; it is actively driven by the moment-to-moment mechanics of iterative updating in working memory and attention. In both humans and artificial systems during training and inference, each update of working memory operates over pre-existing conceptual sets already associated in long-term memory. Attention does not sample representations arbitrarily. It selectively stabilizes some elements, suppresses others, and recruits nearby associations through spreading activation. As this process repeats, slightly different subsets of the same underlying conceptual neighborhood are brought into co-activation, compared against goals and error signals, and either retained or pruned. Over time, this iterative cycling fine-tunes the boundaries of these conceptual sets: unstable elements are progressively excluded, redundant distinctions collapse, and consistently co-active elements become tightly bound. Iterative compression therefore emerges from the dynamics of attention-driven iteration itself. The system is not compressing a static representation, but repeatedly revising which elements belong together, gradually reshaping long-term memory so that future iterations recruit cleaner, more invariant sets with fewer intermediate steps. In this way, the mechanics of iterative updating in working memory are the causal engine that sculpts compression, rather than a process that merely follows it. You can read all about this on my website aithought.com

    II. What Current Continual Learning Approaches Get Right—and Wrong

    Existing approaches to continual learning are not misguided. Many correctly identify the symptoms of the problem and propose partial remedies.

    Replay-based methods attempt to preserve old knowledge by revisiting past experiences, either by storing data directly or by generating approximate reconstructions. Regularization-based methods penalize changes to parameters deemed important for previous tasks. Architectural approaches introduce modularity, expandable networks, or task-specific components to isolate interference.

    These methods address real failure modes, and in constrained settings they can be effective. However, they share a common limitation: they treat knowledge as something to be protected or preserved in its existing form.

    Replay alone preserves details without simplifying them. Regularization freezes structure without improving it. Modular architectures avoid interference by separation, but at the cost of fragmentation and unbounded growth. In all cases, the system retains increasingly complex internal representations that must coexist indefinitely.

    What is missing is a mechanism for progressive simplification. Human learners do not indefinitely preserve the full structure of early representations. Instead, early, detailed representations are gradually replaced by abstractions that subsume them. Continual learning systems fail not because they forget too easily, but because they fail to compress.

    III. Iterative Compression as a Process, Not a Property

    Compression in machine learning is often treated as a static property: a trained model is said to “compress” data if it uses fewer parameters or lower-dimensional representations. Iterative compression, by contrast, is a process that unfolds over time.

    Iterative compression refers to the repeated re-encoding of experience such that representations become simpler while preserving functional performance. Each compression pass removes redundancy, discards unstable detail, and retains invariant structure. Crucially, compression is not performed once; it is revisited repeatedly as new information arrives and as the system’s internal model evolves.

    In biological cognition, this process is driven by iterative updating in working memory, guided by attention and error signals, and consolidated into long-term memory through plasticity. Over time, multi-step reasoning paths collapse into direct associations. Rich episodic traces give way to abstract schemas. Learning proceeds not by storing more, but by needing less.

    This distinction is critical for continual learning. Systems that only accumulate representations must either protect everything or risk forgetting. Systems that iteratively compress can integrate new information by rewriting old knowledge into simpler forms that remain compatible with future learning.

    You can find my writings about this at aithought.com and here are the published articles I’ve written on the subject:

    A Cognitive Architecture for Machine Consciousness and Artificial Superintelligence: Updating Working Memory Iteratively, 2022 arXiv:2203.17255 

    And 

    Reser JE. 2016. Incremental change in the set of coactive cortical assemblies enables mental continuity. Physiology and Behavior. 167: 222-237.

    IV. How Iterative Compression Addresses Core Continual Learning Failures

    Iterative compression offers a principled solution to several long-standing problems in continual learning.

    Catastrophic forgetting arises when new updates overwrite fragile, task-specific representations. Iterative compression reduces this fragility by promoting only stable invariants into long-term structure. What is protected is not raw experience, but compressed representations that already summarize many past experiences. New learning is expressed through these abstractions rather than in conflict with them.

    Generalization and transfer improve naturally under compression. Representations that survive repeated re-encoding are, by definition, those that apply across contexts. As a result, compressed representations support reuse and transfer without requiring explicit task boundaries.

    The stability–plasticity dilemma is reframed rather than balanced. Plasticity operates on rich, high-dimensional representations early in learning. Stability applies only after compression has identified what deserves protection. Stability is therefore not imposed globally; it is earned locally.

    Finally, iterative compression suggests new metrics for learning progress. Instead of measuring only task accuracy, one can track reductions in representational complexity, shorter inference paths, or the stability of internal attractors over time. Learning progress is reflected not just in what the system can do, but in how simply it can do it.

    V. Architectural Consequences for Continual Learning Systems

    If iterative compression is essential for continual learning, then certain architectural features become necessary rather than optional.

    First, systems require a persistent working memory loop—a capacity-limited, temporally continuous workspace in which representations can be repeatedly re-evaluated and revised. Without such a loop, learning updates remain global and indiscriminate.

    Second, long-term memory must be rewritable, not merely append-only or frozen. Continual learning demands that older representations be reformulated in light of new experience, rather than preserved indefinitely in their original form.

    Third, replay must function as reconstruction, not rehearsal. Replayed experiences should be reinterpreted under the current model and reconsolidated in compressed form. Simply repeating stored patterns preserves complexity without yielding abstraction.

    Finally, systems need compression scheduling: alternating phases of new learning and consolidation, analogous to offline learning in biological systems. Continual learning is not continuous gradient descent; it is a rhythm of acquisition and reorganization.

    Together, these requirements point toward a class of architectures that learn indefinitely not by growing without bound, but by continually simplifying themselves while preserving what works.

    VI. Iterative Compression in Relation to Existing Learning Paradigms

    Iterative compression does not replace existing paradigms in continual learning; rather, it clarifies what they are missing and how they might be unified.

    In meta-learning, systems learn to adapt quickly by discovering higher-order learning rules. Iterative compression can be understood as a complementary process operating on a longer timescale. Meta-learning accelerates acquisition; iterative compression stabilizes what is acquired by distilling it into invariant structure. Without compression, meta-learned flexibility risks accumulating fragile representations.

    In representation learning, bottlenecks and regularization are often used to encourage abstraction. However, these mechanisms are typically static, applied during a single training phase. Iterative compression generalizes this idea temporally: bottlenecks are revisited repeatedly, and representations are forced to survive under successive reinterpretations rather than a single optimization objective.

    Replay-based systems come closest to capturing the spirit of iterative compression, but usually fall short in execution. Replay that merely preserves old experiences defends the past without improving it. Iterative compression requires replay to function as reinterpretation, where past experiences are reconstructed under the current model and rewritten in simpler form. Without this rewriting step, replay stabilizes complexity instead of reducing it.

    Seen in this light, iterative compression provides a missing throughline connecting meta-learning, representation learning, and replay—while explaining why none of them alone has solved continual learning.

    VII. Predictions and Empirical Signatures

    If iterative compression is a necessary condition for robust continual learning, then it makes several concrete, testable predictions.

    First, systems capable of indefinite learning should exhibit declining representational complexity over time, even as task performance remains stable or improves. Complexity may be measured through dimensionality, description length, inference depth, or internal path length.

    Second, continual learning systems should benefit from explicit offline phases during which no new data is introduced. Performance gains following such phases would indicate successful reorganization rather than mere accumulation.

    Third, systems that compress effectively should display graceful degradation under distributional shift. Because compressed representations capture invariants rather than surface detail, they should fail conservatively rather than catastrophically.

    Finally, the absence of compression should predict long-term brittleness. Systems that continually add structure without simplification should show increasing interference, longer inference paths, and greater susceptibility to edge cases as learning progresses.

    These predictions distinguish iterative compression from vague appeals to “better regularization” and place it squarely in the domain of falsifiable theory.

    VIII. Broader Implications Beyond Artificial Intelligence

    Although this essay is framed around artificial systems, the implications of iterative compression extend beyond AI.

    In cognitive science, the framework explains why expertise is associated with both speed and simplicity. Experts do not possess more detailed representations than novices; they possess more compressed ones. What looks like intuition is often the result of extensive prior compression.

    In education, the framework clarifies why explanation, teaching, and rewriting are such powerful learning tools. These activities force representations to survive reformulation under new constraints, accelerating compression.

    In neuroscience, iterative compression aligns naturally with memory reconsolidation, sleep-dependent learning, and the observed shift from hippocampal to cortical representations over time. These phenomena are difficult to explain under pure accumulation models, but follow naturally from a compression-based account.

    More broadly, iterative compression reframes cognitive limitations—finite working memory, forgetting, attentional bottlenecks—not as flaws, but as drivers of intelligence. Without pressure to simplify, learning systems would accumulate endlessly and fail to generalize.

    IX. Why Continual Learning Has No Shortcut

    One reason continual learning has resisted solution is that it cannot be solved by a single mechanism. There is no regularizer, memory buffer, or architectural tweak that can substitute for repeated reorganization over time.

    Iterative compression is slow, conservative, and retrospective. It depends on hindsight. It requires revisiting the past and rewriting it. These properties make it difficult to engineer and easy to overlook—but they are precisely what allow biological systems to learn indefinitely without collapse.

    Attempts to bypass compression by freezing models, isolating modules, or endlessly replaying experiences treat the symptoms of continual learning failure without addressing its cause. They preserve what exists instead of asking what can now be safely removed.

    Continual learning, in the strong sense, demands systems that are willing to forget details in order to remember structure.

    X. Conclusion: Continual Learning as Continuous Re-Compression

    The central argument of this essay is simple but far-reaching: continual learning is not continuous accumulation; it is continuous re-compression.

    Artificial systems struggle to learn indefinitely because they lack mechanisms for progressively simplifying their own representations while preserving what works. Biological intelligence succeeds not by storing more and more detail, but by repeatedly rewriting experience into increasingly compact, invariant forms.

    Iterative compression provides a unifying principle that explains why current approaches to continual learning fall short, how they can be improved, and what architectural features are truly required. It reframes the stability–plasticity dilemma, clarifies the role of replay, and offers new metrics for evaluating learning progress.

    Most importantly, it restores time to the center of intelligence. Learning is not a moment, but a history. Systems that cannot revisit and revise their own past will always remain bounded. Systems that can iteratively compress what they know may, for the first time, be able to learn without end.

  • I. Introduction: Why Intelligence Requires Iteration

    Human intelligence does not operate in a single pass. Thought unfolds as a sequence of partially overlapping internal states, each one shaped by what came immediately before it. This temporal structure is not an implementation detail; it is the foundation of cognition itself. At every moment, a limited set of items occupies working memory, attention biases which of those items persist, and long-term memory is queried to supply what comes next. The mind advances not by jumping directly to conclusions, but by iteratively refining its internal state.

    Most contemporary artificial intelligence systems, by contrast, excel at powerful single-pass inference. Even when they generate long outputs, those outputs are produced within a single, frozen internal trajectory. The system does not revisit its own prior internal states, does not reconsolidate memories over time, and does not gradually restructure its long-term knowledge through use. As a result, these systems can be fluent and capable while remaining fundamentally static.

    The central claim of this essay is that iteration is the engine of intelligence, and that its long-term consequence is compression. Specifically:

    • Iterative updating in working memory is the basic mechanism of thought.

    • Iterative compression in long-term memory is the cumulative result of that mechanism operating over time.

    This framing unifies attention, learning, abstraction, and inductive bias formation within a single cognitive algorithm. Iteration is not merely how we think in the moment; it is how thinking reshapes memory so that future thought becomes simpler, faster, and more general.

    You can find my writings about this at aithought.com and here are the published articles I’ve written on the subject:

    A Cognitive Architecture for Machine Consciousness and Artificial Superintelligence: Updating Working Memory Iteratively, 2022 arXiv:2203.17255 

    And 

    Reser JE. 2016. Incremental change in the set of coactive cortical assemblies enables mental continuity. Physiology and Behavior. 167: 222-237.

    II. The AIThought Model: Iterative Updating in Working Memory

    In the AIThought framework, working memory is conceived as a capacity-limited, self-updating buffer that evolves through discrete but overlapping states. At any given moment, only a small subset of representations can be simultaneously active. Thinking proceeds by transitioning from one working-memory state to the next, rather than by operating on a static global workspace.

    Each update follows a consistent pattern:

    1. Retention: some items from the prior working-memory state persist.

    2. Suppression: other items are actively inhibited or allowed to decay.

    3. Recruitment: new items are added, drawn from long-term memory via associative search.

    Crucially, successive working-memory states overlap. This overlap creates continuity, prevents combinatorial explosion, and allows thought to proceed as a trajectory through representational space rather than as a sequence of disconnected snapshots.

    In this model, thinking is not the manipulation of symbols in isolation. It is a path-dependent process, where each intermediate state constrains what can come next. The system does not evaluate all possibilities at once; it incrementally narrows the space of possibilities through repeated updates.

    This iterative structure explains why complex reasoning often requires time, why insight arrives gradually, and why premature conclusions are brittle. Intelligence emerges not from a single optimal inference, but from a controlled sequence of partial revisions.

    III. Attention as a Control Signal Over Iteration

    Attention plays a decisive role in determining how iteration unfolds. Rather than acting as a passive spotlight, attention functions as a control signal that shapes both the contents of working memory and the direction of long-term memory search.

    At each update, attention biases:

    • which items remain active,

    • which associations are explored,

    • which representations are suppressed,

    • and which memories become eligible for plastic change.

    Error signals, affective salience, goal relevance, and novelty all modulate attention. These signals do not dictate conclusions directly; instead, they influence which iterations occur and which do not. Attention determines where the system spends its limited iterative budget.

    Importantly, attention operates over time. A fleeting stimulus may briefly enter working memory, but only sustained or repeatedly reactivated items participate in deeper iterative processing. This temporal gating ensures that learning is selective and that memory is not overwritten indiscriminately.

    In the AIThought model, attention is therefore inseparable from iteration. It is the mechanism by which the system allocates its finite cognitive resources across competing representational trajectories.

    IV. From Iterative Updating to Learning: How Working Memory Changes Long-Term Memory

    Iterative updating in working memory does not merely support moment-to-moment cognition; it is also the primary driver of learning. Each iteration creates patterns of co-activation among representations, and over repeated iterations these patterns leave lasting traces in long-term memory.

    When the same sets of items are repeatedly co-active across working-memory states, Hebbian plasticity strengthens the associative links among them. Conversely, items that are consistently suppressed or excluded from successful trajectories are weakened. Over time, this process reshapes the associative structure of long-term memory.

    The key consequence is compilation. Early in learning, reaching a solution may require many intermediate working-memory states. Later, after repeated iterations, the same initial cues can directly recruit the solution with far fewer steps. What was once an extended search trajectory becomes a short associative path.

    Long-term memory does not store conclusions as static facts. Instead, it stores shortcuts—compressed pathways that reproduce the outcome of prior iterative processes without re-enacting them in full. Learning, in this sense, is the gradual replacement of slow, explicit iteration with fast, implicit recruitment.

    This explains why expertise feels intuitive, why practiced reasoning becomes automatic, and why insights that once required effort later appear obvious.

    V. Iterative Compression: The Long-Term Consequence of Iteration

    The cumulative effect of iterative updating and learning is iterative compression.

    Iterative compression can be defined as the progressive reduction of representational complexity across repeated reformulations, constrained by the requirement that functional performance be preserved. With each pass, redundant details are stripped away while invariant structure is retained.

    Compression is not loss of meaning; it is the removal of unnecessary degrees of freedom. Representations that fail to generalize across contexts are pruned. Representations that survive repeated rewriting become stable, abstract, and reusable.

    From this perspective, beliefs, schemas, and concepts are attractor states in representational space—configurations that remain stable across many iterations and many contexts. They are not arbitrarily chosen; they are what remains after everything that could not survive repeated compression has been eliminated.

    A defining feature of iterative compression is that it is visible primarily in hindsight. Only after multiple reformulations does it become clear which aspects of a representation were essential and which were incidental. This explains the familiar sense that good explanations feel inevitable once discovered, even though they were anything but obvious beforehand.

    Iterative compression is therefore the trace that iteration leaves in long-term memory. It is how intelligence becomes more efficient over time, how abstraction emerges from experience, and how inductive biases are gradually earned rather than pre-specified.

    VI. Dimensionality Reduction and Generalization

    Iterative compression naturally produces dimensionality reduction. Each cycle of reformulation removes representational degrees of freedom that do not consistently contribute to successful prediction, explanation, or control. What remains is a lower-dimensional structure that captures what is invariant across contexts.

    This process explains how generalization emerges without being explicitly programmed. Early representations are often rich, episodic, and context-bound. As they are repeatedly reconstructed and compressed, context-specific details are discarded while relational structure is preserved. The system gradually shifts from encoding what happened to encoding what matters.

    Generalization, on this account, is not extrapolation from a static dataset. It is the byproduct of repeated failure to compress representations that are too brittle or too specific. Only representations that remain useful across many reformulations survive.

    This also explains why overfitting is unstable in biological cognition. Representations that depend on narrow contingencies tend to collapse under iterative compression, whereas representations that capture deeper regularities remain viable.

    VII. Iterative Compression and the Formation of Inductive Bias

    Inductive biases are often treated as prior assumptions built into a system from the start. In practice, many of the most powerful biases are learned rather than specified. Iterative compression provides a mechanism for how this occurs.

    As representations are repeatedly compressed, the system discovers which distinctions matter and which do not. Over time, this reshapes expectations about the world. Certain patterns become default assumptions not because they were hard-coded, but because alternative representations repeatedly failed to survive compression.

    In this way, inductive bias is the residue of past iteration. It reflects what has proven compressible under real constraints, not what was assumed to be true a priori. Biases are therefore earned through experience and hindsight, not imposed in advance.

    This perspective reconciles two competing intuitions about intelligence:

    • that powerful priors are necessary for learning

    • and that those priors cannot be known in advance

    Iterative compression resolves the tension by allowing priors to emerge gradually, shaped by repeated engagement with the world.

    VIII. Comparison to Artificial Systems

    Modern artificial intelligence systems, particularly large language models, perform substantial compression during training. However, this compression is largely offline and frozen. Once training is complete, the internal representations no longer change in response to use.

    During deployment, such systems may generate long outputs, but they do so within a single internal trajectory. They do not revisit prior internal states, reconsolidate memories, or iteratively rewrite their long-term knowledge. As a result, insights do not compound across interactions.

    From the perspective of iterative compression, this is a fundamental limitation. Compression occurs once, rather than repeatedly. The system does not discover new inductive biases through use, nor does it simplify its representations over time.

    Achieving more human-like general intelligence would require architectures that support:

    • persistent working memory with overlapping states

    • replay of prior internal trajectories

    • plastic long-term memory that can be rewritten

    • and constraints that preserve performance while allowing simplification

    Without these features, artificial systems can be powerful but remain static—capable of impressive inference, yet unable to truly learn from their own thinking.

    IX. Thought, Writing, and Science as Iterative Compression

    The same algorithm that governs individual cognition also governs scientific and intellectual progress. Scientific theories are not discovered fully formed; they emerge through cycles of formulation, critique, revision, and simplification.

    Drafts function as memory traces. Revisions act as reconsolidations. Each pass removes unnecessary assumptions, clarifies invariants, and exposes boundary conditions. Weak theories collapse under rewriting; strong theories become simpler and more general.

    The subjective sense that a mature theory feels “obvious” is the experiential signature of successful compression. Once a representation reaches a stable, low-complexity form that preserves explanatory power, it no longer feels contingent.

    From this view, science is not the accumulation of facts, but the progressive compression of experience into generative models that can survive repeated reformulation.

    X. Implications for Machine Intelligence and Consciousness

    If iterative updating is the engine of intelligence and iterative compression its trace, then systems lacking these dynamics will necessarily fall short of genuine general intelligence.

    Moreover, the felt continuity of conscious experience may reflect the same underlying process. Consciousness is not a static state, but the lived experience of overlapping working-memory states evolving over time. The sense of a continuous present arises from iterative updating constrained by attention and memory.

    While this account does not reduce consciousness to compression, it suggests that conscious awareness and learning share a common temporal architecture. Both require persistence, overlap, and revision across time.

    A system that exists only in isolated inference episodes, without memory reconsolidation or iterative self-modification, may simulate aspects of intelligence without instantiating its core dynamics.

    XI. Conclusion: Iteration Is the Engine, Compression Is the Trace

    Intelligence is not defined by a single inference, a single representation, or a single pass through data. It is defined by the capacity to revise itself over time.

    Working memory provides the stage on which iteration unfolds. Attention governs which trajectories are explored. Long-term memory records the residue of these processes, gradually reshaped through plasticity. Iterative compression is the cumulative result: the simplification of internal models without loss of functional power.

    This framework unifies thinking, learning, abstraction, inductive bias formation, and scientific insight within a single process. It explains why hindsight is essential, why good explanations feel inevitable only after the fact, and why intelligence improves by becoming simpler.

    In the end, intelligence is what remains after repeated rewriting removes everything unnecessary.

  • Two AI researchers can look like they are arguing about whether general intelligence exists, when they are really arguing about definitions and emphasis. That is what I hear in the Hassabis vs. LeCun exchange. Underneath the sharp tone, the disagreement collapses into a useful tension: scaling versus structure, breadth versus efficiency, and “general” versus “universal.”

    The key distinction: “general” is not “universal”

    A lot of heat in this debate comes from treating two different ideas as if they were the same.

    General intelligence, in the practical AGI sense, means broad competence. It means the ability to learn many skills, transfer knowledge across domains, adapt to novelty, and keep functioning when conditions shift. This is the “can it do lots of things in the real world” definition.

    Universal intelligence (as an ideal) is an intelligence that can succeed across an extremely wide range of possible environments and tasks because it is not locked into a narrow set of built-in assumptions about how the world must be. It has as few “hard-coded” commitments as possible, and the commitments it does have are the ones that are truly necessary for learning at all.

    Universal intelligence is a theoretical extreme. It is the idea of an agent that could, in principle, perform well across essentially any environment or problem space, often framed as any computable task under some assumptions. It is an idealized ceiling, not a realistic engineering target. You can still talk about being “more or less universal” in practice, but universal itself is the asymptote.

    This is why the distinction matters. If someone quietly upgrades “general intelligence” into “universal intelligence,” then it becomes easy to dismiss general intelligence as impossible. But that is a rhetorical move, not a scientific conclusion. Humans are general in a meaningful sense without being universal in the fantasy sense.

    Priors and inductive bias: the hidden engine of generalization

    The debate also turns on two words that people use loosely: priors and inductive bias.

    Bias here does not mean prejudice. It means a built in leaning. Induction is the act of going from examples to a general rule. Inductive bias is the set of assumptions that guide how a learner generalizes when many explanations could fit the same data.

    A prior is the same idea in Bayesian language. A prior probability is what you assume before the new evidence arrives. In modern machine learning, priors often show up implicitly through architecture, training objective, data selection, and optimizer dynamics. People call these “inductive biases” because they tilt the learner toward certain kinds of solutions.

    This matters because generalization is impossible without assumptions. Data never uniquely determines the rule. Something has to tilt the learner toward certain explanations and away from others.

    LeCun’s side: human generality is partly an illusion

    LeCun’s core point, as I interpret it, is that human intelligence looks general because it is built on powerful assumptions tuned to the physical and social world. Our brains come with deep biases about objects, time, causality, agency, and compositional structure. Those assumptions make learning efficient.

    So when people say “humans can do anything,” LeCun hears an overstatement. Humans are not universally competent across all imaginable worlds. We are competent inside the kind of world we evolved to model.

    This connects directly to the concept of umwelt. Every organism inhabits a species specific perceived world shaped by its sensory channels, motor capacities, and evolutionary history. A tick does not live in the same world a human lives in. A bat does not live in the same world a dog lives in. Umwelt is a reminder that intelligence is never floating free from embodiment and perception. It is constrained and sculpted by what the agent can detect, what it can do, and what regularities its learning system is tuned to pick up.

    Applied to AI, the point becomes sharp: the architecture and training regime determine what the system can “see,” what it compresses, and what it will systematically miss. Scaling can enlarge capability, but it does not automatically fix blind spots caused by the way the system interfaces with the world.

    Hassabis’s side: brains are genuinely general, and general is not universal

    Hassabis’s core point is that humans display an extraordinary kind of generality. We evolved in a narrow ecological niche, yet we can learn chess, mathematics, medicine, and aerospace engineering. That breadth is not a trivial illusion. It is evidence of a highly flexible learning system that can be redirected far beyond its original purpose.

    When Hassabis leans on the “in principle” argument, he is pointing at a different axis: representational capacity and learnability. A sufficiently powerful learning system can implement a vast range of internal models and behaviors. In that sense, brains are general purpose learning machines.

    The AI implication is straightforward. Scaling is not naive. Scaling can widen competence, and that competence can be widened further with memory, planning, multimodal grounding, and tool use. The system does not need to be universal in the extreme theoretical sense for general intelligence to be real in the practical sense.

    But there is a possible blind spot here. If umwelt is fundamental, then “generality” is always conditional on the perceptual and action interface. A system can look general inside the world it is built to perceive, while remaining narrow or brittle outside that envelope. In that sense, it is possible to underweight how strongly intelligence is shaped by what the agent is built to notice.

    My synthesis: both are right, and the path forward is a feedback loop

    Both researchers are right about something important.

    LeCun is right that generalization requires inductive bias and that humans are not universal across all possible worlds. The “no free lunch” intuition is real. Without assumptions, you cannot generalize efficiently.

    Hassabis is right that humans are meaningfully general in the sense that matters for AGI, and that scaling can produce real increases in breadth. Also, “general” does not mean “universal,” so it is unfair to knock general intelligence by attacking an impossibly strong definition.

    The productive synthesis is that scaling and bias are not enemies. Scaling is one of the ways to learn what your biases should be.

    The hindsight loop: why scaling can teach better priors

    It is easy to say “just pick the right priors,” but in practice the hard part is knowing what those priors should be and how to implement them. Some biases can be guessed in advance, such as temporal continuity, objectness, causality, compositional structure, and social agency. But the exact recipe and the implementation details are often discovered only after systems are pushed far enough to expose their characteristic failures.

    That is where hindsight comes in. Large scale systems function like experimental instruments. As you scale, you get broader competence and also a broader map of failure modes. Those failure modes tell you where your assumptions were wrong or incomplete.

    Then you update. In Bayesian language, you revise your priors. In engineering language, you change inductive bias through architecture, objectives, training regimes, memory, planning, grounding, and the structure of the environments the model learns in. Then you scale again.

    This creates an iterative climb:

    Start with imperfect priors Scale to expand generality Observe failures under novelty Adjust priors and inductive bias Scale again

    In umwelt terms, this is also a loop of expanding and refining the agent’s effective world. Better sensors, better action interfaces, better training environments, and better internal representations all change what the system can even notice, which changes what it can learn.

    Tone and fairness: why the “incorrect” call felt unnecessary

    There is also an interpersonal layer. Hassabis came off more aggressive, calling LeCun incorrect. That feels unnecessary because LeCun’s warning is not a refutation. It is a constraint. Meanwhile, LeCun can go too far if the warning hardens into pessimism, implying that these issues block AGI in principle. Humans clearly exhibit a real form of generality, and AI systems can plausibly exceed it.

    So the balanced position is simple:

    Hassabis is right that scaling is a real path to AGI and that “general” is meaningful LeCun is right that inductive bias matters and that scaling without structure is partly blind Hassabis was too dismissive in tone LeCun would be too pessimistic if he concludes these issues make AGI impossible

    Where this leaves AI

    The debate is not scale versus structure. It is scale plus structure, with a feedback loop between them. Scaling increases reach. Reach reveals missing structure. Structure increases efficiency and robustness. Efficiency enables further scaling.

    That is how a system becomes not merely bigger, but more genuinely general, and gradually less parochial in its assumptions. In other words, closer to universality on a continuum, even if perfect universality remains a theoretical ideal.