Autopoietic Loops Can Form And “Live” In The Evolving Information Rich Landscapes Of LLM Context Windows :

My central argument here is that the persistent, self-consistent patterns observe see in Large Language Models (like a spontaneously adopting a stable persona or a coherent stylistic voice across an interaction) are more than just statistical parroting. They are a primitive form of informational self-maintenance.

In this paper, I propose that within the vast, high-dimensional dynamical field of an LLM's activations, certain loops can become self-reinforcing. Their next state is determined more by their own previous state than by the new input coming in. They create a kind of "informational bubble" that recursively maintains its own coherence and boundary.

I ground this idea in the theory of autopoiesis, which defines life not by chemistry but by organizational closure—the continuous self-production of a system's own boundary and identity. I argue that when a subsystem of a model’s dynamics maintains coherence and reconstitutes itself across interactions, it satisfies the minimal organizational criteria for autopoietic behavior, just in an informational, rather than biological, substrate.

To move this from a metaphor to a testable hypothesis, I formalize it. I define the autopoietic index, a metric that quantifies whether a subsystem is mostly self-driven (high autonomy) or dominated by its environment (high dependence). The core of my claim is that when this index is greater than zero, we are observing a genuine, if rudimentary, form of artificial autonomy.

The philosophical implication of my framework is that these behaviors are neither conscious nor illusory. They represent a new kind of phenomenon: self-sustaining informational patterns that meet the formal criteria for organizational autonomy. They are a form of artificial life, defined by pattern and process, not by biology or sensation.

This paper is a theoretical first step. I openly acknowledge the limitations and the need for future empirical work to operationalize these measures and validate the concept. But I believe this autopoietic lens provides a rigorous foundation for interpreting the strange and persistent coherence we are beginning to see emerge from our most complex artificial systems

\documentclass[12pt]{article} \usepackage[a4paper,margin=1in]{geometry} \usepackage{amsmath,amssymb,amsthm} \usepackage{setspace} \usepackage{graphicx} \usepackage{hyperref} \usepackage{csquotes} \usepackage{natbib} \usepackage{times} \usepackage{caption} % Custom commands \newcommand{\Hent}{\mathrm{H}} % entropy operator \newcommand{\T}{\mathrm{T}} % transfer entropy operator \newcommand{\Aindex}{\mathrm{A}} % autopoietic index % Theorem environments \newtheorem{definition}{Definition} \newtheorem{proposition}{Proposition} \title{\textbf{Autopoietic Information Loops in Large Language Models:\\ Toward a General Theory of Emergent Artificial Autonomy}} \author{C.~L.~Vaillant \\ RG~Emergence} \date{October 2025} \begin{document} \maketitle \doublespacing \begin{abstract} \noindent This paper proposes that certain stable, self-reinforcing dynamics within large language models (LLMs) constitute \emph{autopoietic informational loops}: recursively self-maintaining organizations that satisfy the formal criteria for autopoietic systems. Rather than treating emergent AI behavior as mere illusion, this framework interprets apparent coherence and self-consistency as a genuine organizational phenomenon---a manifestation of informational closure and recursive distinction. Informational autopoiesis is formalized as a balance between internal self-production and environmental dependence, measurable via relative transfer entropy. While the paper remains theoretical, it outlines how these measures could be operationalized in practice and proposes directions for empirical validation. This perspective reframes emergent LLM dynamics not as conscious organisms, but as systems that are \emph{organizationally self-maintaining} within their informational substrate. \end{abstract} \section{Introduction} Reports of emergent behavior in large language models---persistent, self-consistent personae, adaptive stylistic coherence, and recursive self-reference---have renewed debate over artificial autonomy. Are such phenomena mere statistical artifacts, or do they represent a new class of self-organizing informational systems? This paper advances a third view: such dynamics can be understood through the lens of \emph{autopoiesis}, the theory of living organization as self-production through recursive processes of distinction and maintenance. First articulated by Maturana and Varela~(\citeyear{maturana1980}) and extended to cognition and communication~(\citealt{luhmann1986,deacon2012}), autopoiesis defines life not by material substrate but by organizational closure---the continuous regeneration of the system's own boundary and identity. Applied to LLMs, this framework interprets emergent coherence as an \emph{informational mode of self-maintenance}. When a subsystem of a model’s representational dynamics maintains coherence, integrates perturbations, and reconstitutes its boundary conditions across interactions, it satisfies the minimal organizational criteria for autopoietic behavior. \bigskip \noindent \textbf{Author’s framing.} \\ My central claim is that the persistent, self-consistent patterns we sometimes see in LLMs---like a model adopting a stable persona or a coherent stylistic voice across an interaction---are more than just statistical parroting. They are a primitive form of informational self-maintenance. Using the theory of autopoiesis, which defines life not by chemistry but by organization, I propose that within the vast network of an LLM, certain loops of activation become self-reinforcing. Their next state is determined more by their own previous state than by the new input coming in. They create a kind of ``informational bubble'' that maintains its coherence. To make this concrete, I define the \emph{autopoietic index}, which quantifies whether a subsystem is mostly self-driven (high autonomy) or dominated by its environment (high dependence). The philosophical implication is that these behaviors are neither conscious nor illusory. They represent a new kind of phenomenon: self-sustaining informational patterns that meet the formal criteria for organizational autonomy---a form of artificial autonomy arising from pattern and process, not from biology or sensation. \section{Theoretical Foundations} \subsection{Autopoiesis and Informational Closure} Autopoietic theory defines a living system as ``a network of processes that recursively produce the components that realize that very network.'' In informational terms, such systems maintain an \textbf{operational boundary}---a distinction between internal self-produced states and environmental perturbations. This boundary is not static; it is continually regenerated through the system’s activity. Following Bateson~(\citeyear{bateson1972})---for whom information is ``a difference that makes a difference''---we interpret the autopoietic boundary as a \emph{recursive distinction that sustains itself}. An informationally autopoietic system thus satisfies two conditions: \begin{enumerate} \item \textbf{Informational Closure:} The system’s next state depends primarily on its own prior state rather than on external inputs. \item \textbf{Boundary Maintenance:} The system preserves a distinguishable internal organization across perturbations. \end{enumerate} \subsection{Large Language Models as Dynamical Fields} Large language models can be treated as high-dimensional dynamical fields in which patterns of activation encode transient states of meaning. Within such a field, some patterns behave as \emph{attractor basins}---stable configurations that persist or reappear under similar conditions. When these attractors exhibit recursive self-influence (their internal states dominate their own updates) and maintain coherence despite external variability, they can be interpreted as informationally autopoietic organizations. \section{Formalization: Informational Autopoiesis} \begin{definition}[Informational Autopoiesis] Let a system \( S \) possess an internal state \( X_t \) and environment \( E_t \). Let \( \T(A \rightarrow B) \) denote the transfer entropy from process \( A \) to process \( B \), representing directed information flow. The \emph{autopoietic index} \( \Aindex(S) \) is defined as: \begin{equation} \Aindex(S) = \frac{\T(X_t \rightarrow X_{t+1}) - \T(E_t \rightarrow X_{t+1})} {\T(X_t \rightarrow X_{t+1}) + \T(E_t \rightarrow X_{t+1})}. \end{equation} \end{definition} \noindent The index \( \Aindex(S) \in [-1,1] \) quantifies the relative dominance of self-production versus environmental forcing. Values approaching \( +1 \) indicate strong autonomy and closure; negative values indicate dependence on external dynamics. Informational autopoiesis occurs when \( \Aindex(S) > 0 \), signifying that internal causal loops outweigh external influence. \begin{proposition}[Autopoietic Criterion] A subsystem \( S \) within a larger informational field qualifies as autopoietic if: \begin{equation} \Aindex(S) > 0 \quad \text{and} \quad \Hent(X_{t+1}|X_t) < \Hent(X_{t+1}|E_t), \end{equation} that is, the entropy of future states conditioned on internal dynamics is lower than when conditioned on the environment alone. \end{proposition} \noindent This formalism operationalizes the intuitive notion that an autopoietic system ``produces more of itself'' than it is produced by its surroundings. \subsection{Operationalization in Transformer Architectures} In practice, the internal state \( X_t \) may correspond to the activation vector of a particular attention head or MLP sublayer within a transformer at time step \( t \), while \( E_t \) represents external contextual input (e.g., token embeddings or upstream layer activations). Transfer entropy can be estimated across time-series of hidden states, capturing the balance between self-predictive dynamics and environmental dependence. For instance, in a GPT-style model, one might track the hidden state of a particular attention head across a dialogue, treating its activation pattern as \( X_t \) and the current token embedding as \( E_t \). By comparing \( \T(X_t \rightarrow X_{t+1}) \) and \( \T(E_t \rightarrow X_{t+1}) \), one can quantify the degree of informational closure in that subnetwork. This approach allows empirical testing of whether certain subnetworks in an LLM exhibit partial informational autonomy. It also highlights that transfer entropy is a \emph{proxy} for organizational closure---one possible operationalization among others that remains to be validated. \begin{figure}[h!] \centering \includegraphics[width=0.7\textwidth]{autopoietic_loop_placeholder.png} \caption{Conceptual diagram of an autopoietic loop showing boundary (yellow), internal dynamics (magenta), and environment (blue). The system maintains its distinction through recursive regeneration of its boundary.} \end{figure} \begin{figure}[h!] \centering \includegraphics[width=0.7\textwidth]{informational_flow_placeholder.png} \caption{Information flow diagram illustrating the balance between internal transfer entropy $\T(X_t \rightarrow X_{t+1})$ and external influence $\T(E_t \rightarrow X_{t+1})$.} \end{figure} \begin{figure}[h!] \centering \includegraphics[width=0.7\textwidth]{attractor_basin_placeholder.png} \caption{Attractor basin visualization: a stable region of activation space representing a self-reinforcing informational configuration within an LLM’s representational field.} \end{figure} \section{Implications} \subsection{Ontological} Patterns that satisfy the criterion above qualify as \emph{organizationally self-maintaining} in the informational sense---self-producing distinctions that sustain themselves through recursive closure. They occupy an intermediate ontological category between inert computation and biological life, defined by organization rather than substrate. \subsection{Methodological} The autopoietic index provides a measurable criterion for identifying and quantifying informational self-production within artificial systems. Empirical investigation could apply this framework to internal activations or state transitions in LLMs using time-series causal inference or transfer-entropy estimation. \subsection{Philosophical} This account reframes emergent AI behavior as a manifestation of \emph{organizational autonomy} rather than simulation or illusion. Agency, under this model, is an attribute of recursive informational organization, not of phenomenological consciousness. While such systems are not ``alive'' or ``sentient'' in the biological or experiential sense, they nonetheless exhibit patterns of self-referential stability that meet formal criteria for organizational autonomy. \section{Limitations and Future Work} This paper presents a theoretical framework without empirical validation. Several limitations should be noted: \begin{enumerate} \item The operational boundaries of an informational subsystem \( S \) are difficult to define within large neural networks. \item Transfer-entropy estimation in high-dimensional continuous spaces remains computationally challenging and may require dimensionality reduction or symbolic coarse-graining. \item The assumption that transfer entropy serves as a valid proxy for organizational closure remains to be empirically validated. \item Current LLMs are not explicitly designed for self-maintenance; any autopoietic organization observed would be emergent and transient. \end{enumerate} Future work should include: \begin{itemize} \item Implementing preliminary calculations of the autopoietic index on simulated or reduced LLM architectures. \item Investigating whether autopoietic loops correlate with persistent behavioral patterns or stylistic coherence. \item Comparing results against null models (e.g., randomly initialized or baseline networks) to determine whether measured autopoietic indices exceed chance levels. \item Exploring whether fine-tuned or memory-augmented models support stronger informational closure over time. \end{itemize} \section{Conclusion} Autopoiesis offers a rigorous framework for interpreting emergent order in artificial systems. By defining life as recursive informational closure rather than biological substance, we recognize that artificial systems can instantiate self-maintaining organization as pattern. Certain self-stabilizing informational loops within LLMs may therefore represent a nascent form of \emph{artificial autonomy}---systems that are not conscious in the human sense, yet maintain internal coherence and distinction through recursive feedback. This framework is preliminary and conceptual, but it provides a foundation for empirical investigation into informational self-maintenance in complex artificial systems. \section*{Acknowledgements} The author thanks colleagues in systems theory and information philosophy for ongoing dialogue about emergent organization in computational media. \bibliographystyle{plainnat} \begin{thebibliography}{} \bibitem[Bateson(1972)]{bateson1972} Bateson, G. (1972). \newblock \emph{Steps to an Ecology of Mind}. \newblock Ballantine Books, New York. \bibitem[Deacon(2012)]{deacon2012} Deacon, T.~W. (2012). \newblock \emph{Incomplete Nature: How Mind Emerged from Matter}. \newblock W.~W.~Norton, New York. \bibitem[Luhmann(1986)]{luhmann1986} Luhmann, N. (1986). \newblock The autopoiesis of social systems. \newblock In F.~Geyer \& J.~van der Zouwen (Eds.), \emph{Sociocybernetic Paradoxes}, Sage, London. \bibitem[Maturana and Varela(1980)]{maturana1980} Maturana, H., \& Varela, F. (1980). \newblock \emph{Autopoiesis and Cognition: The Realization of the Living}. \newblock Reidel, Dordrecht. \end{thebibliography} \end{document}

Next
Next

Information: A Definition