\documentclass[12pt]{article}
% =======================
% Packages
% =======================
\usepackage[a4paper,margin=1in]{geometry}
\usepackage{amsmath,amssymb,amsfonts,amsthm}
\usepackage{setspace}
\usepackage{csquotes}
\usepackage{natbib}
\usepackage{times} % If you prefer better math with Times look, swap for: newtxtext,newtxmath
\usepackage{caption}
% TikZ/PGF for embedded figures (no external images)
\usepackage{tikz}
\usetikzlibrary{arrows.meta,positioning,calc,fit,shapes.geometric,decorations.pathmorphing}
\usepackage{pgfplots}
\pgfplotsset{compat=1.18}
% Keep hyperref last
\usepackage[hidelinks]{hyperref}
% =======================
% Commands & Theorems
% =======================
\newcommand{\Hent}{\mathrm{H}} % entropy operator
\DeclareMathOperator*{\argmin}{arg\,min}
\newtheorem{definition}{Definition}
\newtheorem{proposition}{Proposition}
% =======================
% Metadata
% =======================
\title{\textbf{Information as Distinction: The Foundational Difference that Drives Emergence}}
\author{C.~L.~Vaillant \\ RG~Emergence}
\date{October 2025}
\begin{document}
\maketitle
\doublespacing
\begin{abstract}
\noindent
This paper proposes a general definition of information as \emph{the minimal distinguishable difference that reduces uncertainty within a system}. Information is treated not as a substance but as a relational pattern---an identifiable distinction that can be registered by a system. From this foundation, intelligence, structure, and meaning are understood as emergent properties of recursive informational dynamics. The argument situates this view within the lineage of Shannon's mathematical theory of communication, Bateson's cybernetic epistemology, Wheeler's ``it from bit'' cosmology, and contemporary emergence theory.
\end{abstract}
\section{Introduction}
Information underlies both the physical and cognitive domains. Whether in the spin of an electron, the firing of a neuron, or the choice within a decision space, reality presents itself as a field of possible states. The process by which a distinction is drawn within that field---the recognition of a difference that makes a difference---constitutes the most primitive act of information.
In this sense, information is ontological as well as epistemic. It describes not only how observers come to know, but how being itself differentiates. This view aligns with the broader movement in systems and emergence theory that sees complexity, agency, and meaning as products of recursive informational organization.
\section{Defining Information}
\begin{definition}[General Form]\label{def:info}
Information is the minimal distinguishable difference that reduces uncertainty within a system.
\end{definition}
Information is not an object or a medium but a relation among possible states. It exists only when three conditions obtain:
\begin{enumerate}
\item A \textbf{state space} \( S \): the set of possible configurations of a system.
\item A \textbf{difference} \( \Delta s \): a distinguishable contrast between states in \( S \).
\item An \textbf{observer or subsystem} \( O \): an entity capable of registering or responding to \( \Delta s \).
\end{enumerate}
Formally, we can express this relation as:
\begin{equation}
I(S,\Delta s,O)\ \Longleftrightarrow\ \exists\, \Delta s \in S:\ O \text{ registers } \Delta s \text{ and } \Hent(S)\downarrow,
\end{equation}
where \( \Hent(S) \) denotes the Shannon entropy---a measure of uncertainty over possible states of \( S \).
Thus, information corresponds to the reduction of uncertainty through the recognition of a difference.
% -----------------------
% Figure 1: State space, distinction, and observer (TikZ)
% -----------------------
\begin{figure}[h!]
\centering
\begin{tikzpicture}[>=Latex, thick, node distance=1.2cm]
% Colors
\definecolor{Sblue}{RGB}{40,120,200}
\definecolor{Ogreen}{RGB}{60,150,80}
\definecolor{Dred}{RGB}{190,50,60}
% State space ellipse
\node[ellipse, draw=Sblue!80!black, minimum width=9.5cm, minimum height=5.5cm, label={[Sblue!70!black]90:$S$}] (S) {};
% Two states inside S
\node[circle, draw=Dred!70!black, fill=Dred!10, minimum size=9mm] (s1) at (-1.8,0.5) {$s_1$};
\node[circle, draw=Dred!70!black, fill=Dred!10, minimum size=9mm] (s2) at (0.8,0.2) {$s_2$};
% Delta s arrow
\draw[->, Dred!90!black, line width=1.1pt] (s1) -- node[above] {$\Delta s$} (s2);
% Observer
\node[draw=Ogreen!80!black, fill=Ogreen!10, rounded corners, minimum width=14mm, minimum height=8mm, right=3.0cm of s2] (O) {$O$};
% Registration arrow
\draw[->, Ogreen!80!black, line width=1.0pt] (O.west) -- ++(-1.4,0) node[midway, above] {\footnotesize register};
% Entropy reduction callout
\node[draw, rounded corners, align=center, below=1.1cm of s1] (ent) {$\Hent(S)$ reduces when $O$\\distinguishes $\Delta s$};
\draw[->, dashed] (ent.north) -- ($(s1.south west)!0.5!(s2.south)$);
\end{tikzpicture}
\caption{Information as a registered distinction: an observer $O$ distinguishes $\Delta s$ within the state space $S$, thereby reducing uncertainty $\Hent(S)$.}
\end{figure}
\section{Formalization and Context}
Shannon~(\citeyear{shannon1948}) defined the quantity of information as the expected reduction in uncertainty:
\begin{equation}
I = \Hent(X) - \Hent(X|Y),
\end{equation}
where \( \Hent(X) \) is the entropy of a random variable \( X \) and \( \Hent(X|Y) \) is the conditional entropy after observing \( Y \).
Bateson~(\citeyear{bateson1972}) reformulated this notion qualitatively as ``a difference that makes a difference.'' Our present definition synthesizes these perspectives by grounding both in a minimal formal and phenomenological requirement: information as the smallest distinguishable change capable of reducing uncertainty for a system capable of registering it.
% -----------------------
% Figure 2: Shannon information flow (TikZ)
% -----------------------
\begin{figure}[h!]
\centering
\begin{tikzpicture}[>=Latex, node distance=2.2cm, thick, every node/.style={font=\small}]
\node[draw, rounded corners, minimum width=20mm, minimum height=9mm] (X) {$X$};
\node[draw, rounded corners, minimum width=20mm, minimum height=9mm, above=0.9cm of X] (Y) {$Y$};
\node[draw, rounded corners, minimum width=28mm, minimum height=10mm, right=4.2cm of X] (HX) {$\Hent(X)$};
\node[draw, rounded corners, minimum width=34mm, minimum height=10mm, below=0.0cm of HX, right=4.2cm of X] (HXY) {$\Hent(X|Y)$};
\draw[->] (X) -- node[below] {} (HX);
\draw[->] (Y) -- node[above, sloped] {} (HXY);
\draw[->, dashed] (Y) -- (HX);
\draw[->, line width=1.1pt] (HX) -- node[above] {$I=\Hent(X)-\Hent(X|Y)$} ++(2.7,0);
\draw[->, line width=1.1pt] (HXY) -- ++(2.7,0);
\node[align=center, below=1.0cm of HXY, text width=0.7\textwidth]
{Observation $Y$ reduces the uncertainty about $X$, yielding information $I$.};
\end{tikzpicture}
\caption{Shannon information as uncertainty reduction: $I=\Hent(X)-\Hent(X|Y)$.}
\end{figure}
\section{Recursive Dynamics and Emergence}
\begin{proposition}[Recursive Information Process]
Let \( I_t \) represent the informational state of a system at time \( t \). The recursive dynamic is given by
\begin{equation}
I_{t+1} = f(I_t, \Delta I_t),
\end{equation}
where \( f \) encodes how new distinctions \( \Delta I_t \) are integrated into prior informational structure.
\end{proposition}
Through iteration, such processes yield emergent structures: feedback loops stabilize into attractors, and uncertainty transforms into order.
\subsection{Example: Adaptive Agent}
Consider an adaptive agent learning from an environment \( E \). Each observation introduces a distinction \( \Delta s_t \) that updates the agent’s internal model:
\begin{equation}
M_{t+1} = M_t + g(\Delta s_t),
\end{equation}
where \( g \) represents an assimilation function. Over time, \( M_t \) converges toward a more predictive state, embodying the emergence of intelligence through recursive information refinement.
% -----------------------
% Figure 3: Recursive integration block diagram (TikZ)
% -----------------------
\begin{figure}[h!]
\centering
\begin{tikzpicture}[>=Latex, thick, node distance=1.8cm, auto, every node/.style={font=\small}]
% Nodes
\node[draw, rounded corners, minimum width=20mm, minimum height=9mm] (It) {$I_t$};
\node[circle, draw, minimum size=9mm, right=1.6cm of It] (sum) {$\Sigma$};
\node[draw, rounded corners, minimum width=22mm, minimum height=9mm, right=1.8cm of sum] (f) {$f(\cdot)$};
\node[draw, rounded corners, minimum width=20mm, minimum height=9mm, right=1.8cm of f] (Itp1) {$I_{t+1}$};
% Delta I input
\node[draw, rounded corners, minimum width=20mm, minimum height=9mm, above=1.2cm of sum] (dI) {$\Delta I_t$};
% Connections
\draw[->] (It) -- (sum);
\draw[->] (dI) -- (sum);
\draw[->] (sum) -- (f);
\draw[->] (f) -- (Itp1);
% Feedback loop
\draw[->] (Itp1.east) -- ++(0.8,0) |- node[pos=0.25, right] {\footnotesize integration memory} (f.south);
% Annotation
\node[align=center, below=0.9cm of f, text width=0.78\textwidth]
{New distinctions $\Delta I_t$ are integrated with prior structure $I_t$ to produce $I_{t+1}$; iterating this loop yields emergent order.};
\end{tikzpicture}
\caption{Recursive integration of distinctions drives emergence.}
\end{figure}
\section{Discussion and Counterpoints}
One might object that such a framework collapses into pancomputationalism---the claim that the universe is merely computing itself. However, the distinction-based model avoids this reduction by grounding computation in relational ontology: not all dynamics are informational, only those involving distinguishable and registered differences.
Further, this view aligns with complexity theory’s emphasis on constraint and self-organization rather than universal computation. Information, as distinction, underwrites the generation of new levels of order without presupposing total computability.
\section{Implications}
This definition carries implications across multiple domains:
\begin{itemize}
\item \textbf{Physics:} Wheeler’s aphorism ``it from bit''~\citep{wheeler1990} finds grounding---physical reality arises from networks of distinctions.
\item \textbf{Biology:} Life can be viewed as recursive information maintenance---autopoietic systems preserving distinctions against entropy~\citep{maturana1980,deacon2012}.
\item \textbf{Cognition:} Intelligence becomes the continual refinement of informational structure through learning, compression, and transformation.
\end{itemize}
\section{Conclusion}
Information, defined as the minimal distinguishable difference that reduces uncertainty, provides a unifying foundation for understanding emergence. It is the substrate of structure and the grammar of becoming. From elementary physics to collective intelligence, the world unfolds as a recursive play of differences organizing themselves into meaning.
\bibliographystyle{plainnat}
\begin{thebibliography}{}
\bibitem[Shannon(1948)]{shannon1948}
Shannon, C.~E. (1948).
\newblock A mathematical theory of communication.
\newblock \emph{Bell System Technical Journal}, 27(3), 379--423.
\bibitem[Bateson(1972)]{bateson1972}
Bateson, G. (1972).
\newblock \emph{Steps to an Ecology of Mind}.
\newblock Ballantine Books, New York.
\bibitem[Wheeler(1990)]{wheeler1990}
Wheeler, J.~A. (1990).
\newblock Information, physics, quantum: The search for links.
\newblock In W. Zurek (Ed.), \emph{Complexity, Entropy, and the Physics of Information}, Addison--Wesley.
\bibitem[Maturana and Varela(1980)]{maturana1980}
Maturana, H., \& Varela, F. (1980).
\newblock \emph{Autopoiesis and Cognition: The Realization of the Living}.
\newblock Reidel, Dordrecht.
\bibitem[Deacon(2012)]{deacon2012}
Deacon, T.~W. (2012).
\newblock \emph{Incomplete Nature: How Mind Emerged from Matter}.
\newblock W.~W.~Norton, New York.
\end{thebibliography}
\end{document}
want me to swap the times package for newtxtext,newtxmath or add microtype for nicer spacing and protrusion? I can drop those in cleanly.