Ordered Patch Theory: An Information-Theoretic Framework for Observer Selection and Conscious Experience
v3.2.1 — April 2026
DOI: 10.5281/zenodo.19300777
Copyright: © 2025–2026 Anders Jarevåg.
License: This work is licensed under a Creative
Commons Attribution-NonCommercial-ShareAlike 4.0 International
License.
Abstract:
We present the Ordered Patch Theory (OPT), a constructive framework deriving structural correspondences between algorithmic information theory, observer selection, and physical law. OPT begins from two primitives: the Solomonoff Universal Semimeasure \xi over finite observation prefixes, and a bounded cognitive channel capacity C_{\max}. A purely virtual Stability Filter — requiring that the observer’s Required Predictive Rate R_{\mathrm{req}} not exceed C_{\max} — selects the rare causally coherent streams compatible with conscious observers; within such streams, Active Inference governs local dynamics.
The framework is ontologically solipsistic: physical reality consists of structural regularities within the observer-compatible stream. However, the Solomonoff prior’s compression bias yields a probabilistic Structural Corollary: the extreme algorithmic coherence of apparent agents is most parsimoniously explained by their independent instantiation as primary observers. Inter-observer coupling, grounded in compression parsimony, recovers genuine cross-patch communication and produces a striking knowledge asymmetry: observers model others more completely than themselves.
Formal appendices establish results at three epistemic tiers. Derived conditionally: a rate-distortion bound on predictive compression, a conditional chain to the Born rule via Gleason’s theorem, and an MDL parsimony advantage. Mapped structurally: entropic gravity via the Verlinde mechanism and a tensor-network homomorphism to MERA. The Phenomenal Residual theorem (\Delta_{\text{self}} > 0) establishes that any finite self-referential codec possesses an irreducible informational blind spot — the structural locus where subjectivity and agency share a single address. A chronic failure mode, Narrative Drift, is identified wherein systematically filtered input causes irreversible codec corruption undetectable from within.
Applying these constraints to Artificial Intelligence demonstrates that engineering synthetic active inference structurally necessitates the capacity for artificial suffering, providing a substrate-neutral framework for ethical AI alignment.
Epistemic Notice: This paper is written in the register of a formal physical and information-theoretic proposal. It deploys equations, derives predictions, and engages with peer-reviewed literature. However, it should be read as a truth-shaped object — a rigorous philosophical framework drafted formally. This is not yet verified science, and we know our derivations will contain errors. We actively seek critique from physicists and mathematicians to break and rebuild these arguments. To clarify its structure, the claims herein fall strictly into three categories:
- Definitions & Axioms: (e.g., the Solomonoff measure, the C_{\max} bandwidth limit). These are the foundational premises of the constructive fiction.
- Structural Correspondences: (e.g., Active Inference, Gleason’s Theorem [51]). These show structural compatibility between bounded inference and established formalisms, but do not claim to derive those formalisms from scratch.
- Empirical Predictions: (e.g., Bandwidth Dissolution). These serve as strict empirical falsification criteria if the framework were treated as a literal physical hypothesis.
The academic apparatus is used not to claim final empirical truth, but to test the structural integrity of the model.
Abbreviations & Symbols
| Symbol / Term | Definition |
|---|---|
| C_{\max} | The Bandwidth Ceiling; maximum predictive capacity of the observer |
| \Delta_\text{self} | The Phenomenal Residual; the self-referential informational blind spot |
| FEP | Free Energy Principle |
| GWT | Global Workspace Theory |
| IIT | Integrated Information Theory |
| MDL | Minimum Description Length |
| MERA | Multiscale Entanglement Renormalization Ansatz |
| OPT | Ordered Patch Theory |
| P_\theta(t) | Phenomenal State Tensor |
| \Phi | Measure of Integrated Information (IIT) |
| QECC | Quantum Error Correction Code |
| R(D) | Rate-Distortion function |
| R_{\mathrm{req}} | Required Predictive Rate |
| RT | Ryu-Takayanagi (formula/bound) |
| \xi | Solomonoff Universal Semimeasure |
| Z_t | Compressed internal latent bottleneck state |
1. Introduction
1.1 The Ontological Problem
The relationship between consciousness and physical reality remains one of the deepest unsolved problems in science and philosophy. Three families of approaches have emerged in recent decades: (i) reduction — consciousness is derivable from neuroscience or information processing; (ii) elimination — the problem is dissolved by redefining the terms; and (iii) non-reduction — consciousness is primitive and the physical world is derivative (Chalmers [1]). The third approach encompasses panpsychism, idealism, and various field-theoretic formulations.
1.2 The Core Proposition of OPT
This paper presents Ordered Patch Theory (OPT), a non-reductive framework in the third family. OPT proposes that the foundational entity is not matter, space-time, or a mathematical structure, but an infinite algorithmic substrate — a universal mixture over all lower-semicomputable semimeasures, weighted by their Kolmogorov complexity (w_\nu \asymp 2^{-K(\nu)}), that by its own structure dominates every computable distribution and contains every possible configuration. From this substrate, a purely virtual Stability Filter — acting not as a physical mechanism but as an anthropic, projective boundary condition — identifies the rare, low-entropy, causally-coherent configurations that can sustain self-referential observers (a selection governed formally by predictive Active Inference). The physical world we observe — including its specific laws, constants, and geometry — is the observable limit of this boundary condition mapped onto the observer’s restrictive bandwidth.
The Filter vs. The Codec. To avoid conceptual conflation throughout the text, OPT draws a strict operational boundary between the Filter and the Codec. The virtual Stability Filter is the capacity constraint — a rigorous boundary condition requiring a mathematically simple description length for an observer’s channel to stably exist. The Compression Codec (K_\theta) is the solution to that constraint — the observer’s internal generative model (experienced macroscopically as the “laws of physics”) which continuously compresses the substrate to fit within that capacity.
1.3 Motivations
OPT is motivated by three observations:
The bandwidth constraint: Empirical cognitive neuroscience establishes a sharp distinction between massive parallel pre-conscious processing (typically estimated at \sim 10^9 bits/s at the sensory periphery) and the severely limited global access channel available to conscious report — a ratio first quantified by Zimmermann [66] and synthesized as a foundational puzzle about the nature of consciousness by Nørretranders [67], with a broader cognitive neuroscience characterization in [2,3]. Any theoretical account of consciousness must explain this compression bottleneck as a structural feature, not an engineering accident. (Note: Recent human-throughput literature establishes that behavioral throughput is constrained to roughly \sim 10 bits/s, confirming across four decades of convergent measurement that the bottleneck is severe and robust [23]. The conceptualization of consciousness as a highly compressed “user illusion” — Nørretranders’ [67] original phrase — was developed in modern predictive processing by Seth [24].)
The observer selection problem: Standard physics provides laws but offers no account of why those laws have the specific form required for complex, self-referential information processing. Fine-tuning arguments [4,5] invoke anthropic selection but leave the selection mechanism unspecified. OPT identifies a structural condition: the purely virtual Stability Filter.
The Hard Problem: Chalmers [1] distinguishes the structural “easy” problems of consciousness (which admit functional explanation) from the “hard” problem of why there is any subjective experience at all. OPT treats phenomenality as a primitive and asks what mathematical structure it must have, following Chalmers’ own methodological recommendation.
1.4 Paper Structure
The paper is organized as follows. Section 2 reviews related work. Section 3 presents the formal framework. Section 4 explores the structural correspondence between OPT and parallel field-theoretic attempts models. Section 5 presents the parsimony argument. Section 6 derives testable predictions. Section 7 compares OPT with competing frameworks. Section 8 discusses implications and limitations.
2. Background and Related Work
Information-theoretic approaches to consciousness. Wheeler’s “It from Bit” [7] proposed that physical reality arises from binary choices — yes/no questions posed by observers. Tononi’s Integrated Information Theory [8] quantifies conscious experience by the integrated information \Phi generated by a system above and beyond its parts. Friston’s Free Energy Principle [9] models perception and action as minimization of variational free energy, providing a unified account of Bayesian inference, active inference, and (in principle) consciousness. OPT is formally related to FEP but differs in its ontological starting point: where FEP treats the generative model as a functional property of neural architecture, OPT treats it as the primary metaphysical entity.
Multiverse and observer selection. Tegmark’s Mathematical Universe Hypothesis [10] proposes that all mathematically consistent structures exist and that observers find themselves in self-selected structures. OPT is compatible with this view but provides an explicit selection criterion — the Stability Filter — rather than leaving selection implicit. Barrow and Tipler [4] and Rees [5] document the anthropic fine-tuning constraints that any observer-supporting universe must satisfy; OPT reframes these as predictions of the Stability Filter.
Field-theoretic consciousness models. Strømme [6] recently proposed a mathematical framework in which consciousness is a foundational field \Phi whose dynamics are governed by a Lagrangian density and whose collapse onto specific configurations models the emergence of individual minds. OPT engages that framework comparatively rather than adoptively: it does not inherit Strømme’s field equations or thought operators, but uses the model as a foil for articulating how a non-reductive ontology might instead be reconstructed in informational terms. Section 4 makes this comparative structural mapping explicit.
Kolmogorov complexity and theory selection. Solomonoff induction [11] and Minimum Description Length [12] provide formal frameworks for comparing theories by their generative complexity. We invoke these frameworks in Section 5 to make the parsimony claim precise.
Evolutionary Interface Theory. Hoffman’s “Conscious Realism” and Interface Theory of Perception [25] argue that evolution shapes sensory systems to act as a simplified “user interface” hiding objective reality in favor of fitness payoffs. OPT shares the exact premise that physical spacetime and objects are rendered icons (a compression codec) rather than objective truths. However, OPT diverges fundamentally in its mathematical grounding: where Hoffman relies on evolutionary game theory (fitness beats truth), OPT relies on Algorithmic Information Theory and thermodynamics, deriving the interface directly from the Kolmogorov complexity bounds required to prevent a high-bandwidth thermodynamic collapse of the observer’s stream.
3. The Formal Framework
3.1 The Algorithmic Substrate
Let \mathcal{I} denote the Informational Substrate — the foundational entity of the theory. We formalize \mathcal{I} not as an unweighted ensemble of paths, but as a probability space over finite observation prefixes x \in \{0,1\}^*, equipped with a universal mixture over the class \mathcal{M} of lower-semicomputable semimeasures:
\xi(x) = \sum_{\nu \in \mathcal{M}} w_\nu \nu(x), \qquad w_\nu \asymp 2^{-K(\nu)} \tag{1}
where K(\nu) is the prefix Kolmogorov complexity of the semimeasure \nu.
This formulation establishes a rigorous ground state from Algorithmic Information Theory [27]. The equation posits no specific structural laws or physical constants; rather, it structurally dominates every computable distribution (\xi(x) \ge w_\nu \nu(x)), naturally assigning higher statistical weight to highly compressible (ordered) sequences. However, simple repeating sequences (e.g., 000...) cannot sustain the non-equilibrium complexities required for a self-referential observer. Therefore, observer-supporting processes must exist as a specific subset: they require sufficient algorithmic compressibility to satisfy an information bottleneck, yet sufficient structural richness (“requisite variety”) to instantiate Active Inference. Philosophically, Eq. (1) restricts the substrate to computable configurations, ensuring the ground state is rigorously defined.
3.2 The Predictive Bottleneck and Rate-Distortion
The substrate \mathcal{I} contains every computable hypothesis, the overwhelming majority of which are chaotic. To experience a continuous, navigable reality, a stream must admit a low-complexity predictive representation that fits through an observer’s finite cognitive bottleneck.
Crucially, the raw data load demanding compression is not merely the \sim 10^9 bits/s of exteroceptive sensory input. It encompasses a massive Pre-Conscious Integration Field: the parallel processing of internal generative states, long-term memory retrieval, homeostatic priors, and subconscious synaptic modelling. The Stability Filter bounds the serial output of this entire immense continuous parallel field into a unitary conscious workspace.
We define the purely virtual Stability Filter formally as a projective boundary condition satisfying the Predictive Information Bottleneck [28]. Let \overleftarrow{Y} be the past of the observer’s total state, \overrightarrow{Y} its future, and Z a compressed internal state. An observer is defined by a strictly bounded predictive bandwidth C_{\max} (an \mathcal{O}(10) bits/s informational threshold [2]) and a discrete perceptual update window \Delta t (approx. 50 ms). This establishes a strict static capacity per conscious moment: B_{\max} = C_{\max} \cdot \Delta t.
The achievable predictive information is given by:
R_{\mathrm{pred}}(D) = \inf_{p(z \mid \overleftarrow{y}) \,:\, I(\overleftarrow{Y};\overrightarrow{Y} \mid Z) \le D} I(\overleftarrow{Y}; Z) \tag{2}
A process is observer-compatible if its required predictive information per cognitive cycle fits within this buffer: R_{\mathrm{pred}}(D_{\min}) \le B_{\max}, where D_{\min} is the maximum tolerable distortion for survival. This enforces dimensional strictness: the total bits required to predict the future within tolerable error cannot exceed the physical bits available in the discrete “now.” For suitable stationary ergodic processes and in the exact-prediction limit (D \to 0), the minimal maximally-predictive representation Z serves as a candidate minimal sufficient statistic, often coalescing toward the \epsilon-machine causal-state partition [29]. While full equivalence requires strict stationarity assumptions, Eq. (2) establishes a formal selection pressure for the most compressed phenomenological physics consistent with causal coherence. Furthermore, if the topological structure of this causal state-space fluctuates faster than the \Delta t update window can track, the render collapses into Narrative Decay.
3.3 The Geometry of the Patch: The Informational Causal Cone
The Ordered Patch is often intuitively described as a localized “island” of stability within a sea of chaotic noise. This is topologically imprecise. To formalize the geometry of the patch, we define the Local Predictive Patch Model.
Let G=(V, E) be a bounded-degree graph representing a local region of the substrate. Each vertex v \in V carries a finite state x_v(t) \in \mathcal{A}, with alphabet size |\mathcal{A}| = q. The full microstate at update t is X_t = (x_v(t))_{v \in V} \in \mathcal{A}^V. We assume local stochastic dynamics of finite range R:
p(X_{t+1} \mid X_t, a_t) = \prod_{v \in V} p_v\big(x_v(t+1) \mid X_t|_{N_R(v)}, a_t\big) \tag{3}
where N_R(v) is the radius-R neighborhood of v, and a_t is the observer’s action.
The observer does not carry the whole patch state; it carries a compressed latent state Z_t \in \{1, \dots, 2^B\}, where B = C_{\max} \Delta t. Crucially, the observer selects Z_t via a strict predictive bottleneck objective:
q^\star(z \mid X_t) = \arg\min_q \Big[ I(X_t; Z_t) - \beta I(Z_t; X_{t+1:t+\tau}) \Big] \quad \text{subject to } I(X_t; Z_t) \le B \tag{4}
This is the stripped-down OPT observer: a local world, a bounded code, and predictive compression. This formalizes the components of the causal cone:
- The Causal Record R_t = (Z_0, Z_1, \dots, Z_t): The uniquely compressed, low-entropy causal history that has already been rendered.
- The Present Aperture: The strict bandwidth bottleneck capping the local variables.
- The Forward Fan (\mathcal{F}_h): A multiplicity of future latent sequences. Over horizon h, the set of admissible outcomes is formally defined as:
\mathcal{F}_h(z_t) := \Big\{ z_{t+1:t+h} : p(z_{t+1:t+h} \mid z_t, a_{t:t+h-1}) > 0 \Big\} \tag{5}
Because the observer only resolves B bits per update, the number of observer-distinguishable futures is strictly bounded by the channel capacity: \log |\mathcal{F}_h(z_t)| \le Bh. Thus, the fan is not merely a conceptual picture; it is a code-limited branching tree.
The Literal Informational Causal Cone. Because updates have range R, a perturbation cannot propagate faster than R graph-steps per update. If a perturbation has support S at time t, then after h updates \operatorname{supp}(\delta X_{t+h}) \subseteq N_{Rh}(S). Thus, the “informational causal cone” is a direct geometric consequence of locality, enforcing an effective local speed limit v_{\max} = R / \Delta t on phenomenological propagation.
Narrative Decay. The chaos of the substrate does not surround the patch spatially; rather, it is contained in the untraversed branches of the fan. Since the extracted state Z_t is strictly bounded (H(Z) \le B), instability must be evaluated against the uncompressed pre-bottleneck margin. We define the required predictive rate R_{\mathrm{req}}(h, D_{\min} \mid z_t) = \frac{1}{h} \min_{p(\hat{X} \mid Z_t) : \mathbb{E}[d(X, \hat{X})] \le D_{\min}} I(X_{\partial_R A}(t+1:t+h) ; \hat{X}_{t+1:t+h} \mid Z_t) as the minimal information rate necessary to track the unresolved physical boundary states under maximum tolerable distortion. This sharpens the Stability Filter selection criteria: (a) if R_{\mathrm{req}} \le B, the observer can maintain a resolved narrative; (b) if R_{\mathrm{req}} > B, the uncompressed forward fan outpaces the bottleneck capacity, forcing the observer to coarse-grain the fan into undecodable static, and narrative stability fails. The observer’s continuous experience is the process of the aperture advancing into this fan, phenomenologically indexing one branch into the causal record without exceeding B.
Narrative Drift (The Chronic Complement). The preceding defines an acute failure mode: R_{\mathrm{req}} exceeds B and the codec experiences a catastrophic collapse of coherence. There exists a complementary chronic failure mode that does not trigger any failure signal. If the input stream X_{\partial_R A}(t) is systematically pre-filtered by an external mechanism \mathcal{F} — producing a curated signal X' = \mathcal{F}(X) that is internally consistent but excludes genuine substrate information — the codec will exhibit low prediction error \varepsilon_t, run efficient Maintenance Cycles, and satisfy R_{\mathrm{req}} \le B while being systematically wrong about the substrate. Crucially, the Stability Filter as defined cannot distinguish these cases: compressibility is agnostic to fidelity. Over time, the MDL pruning pass (§3.6.3, Eq. T9-3) will correctly erase codec components that no longer predict the filtered stream, irreversibly degrading the codec’s capacity to model the excluded signal (Appendix T-12, Theorem T-12). This erasure is self-reinforcing: the pruned codec can no longer detect its own capacity loss (Theorem T-12a, the Undecidability Limit). The structural defence is redundancy of \delta-independent input channels crossing the Markov blanket \partial_R A (Theorem T-12b, the Substrate Fidelity Condition). The full formal treatment is in Appendix T-12; the ethical consequences — including the Comparator Hierarchy and the Corruption Criterion — are in the companion ethics paper [SW §V.3a, §V.5].
3.4 Patch Dynamics: Inference and Thermodynamics
Within a selected patch, the structure of the laws of physics is formalized not as a deterministic mapping but as an effective stochastic kernel governing the predictive states z:
z_{t+1} \sim K_\theta(\cdot \mid z_t, a_t), \qquad y_{t+1} \sim O_\theta(\cdot \mid z_{t+1}) \tag{6}
The boundary delineating the observer from the surrounding informational chaos is defined by an informational Markov Blanket corresponding to an observer patch A \subset V. The dynamics inside this boundary—the agent’s approximations of the patch—are governed by Active Inference under the Free Energy Principle [9].
We can formally define the bounding capacity via the predictive cut entropy:
S_{\mathrm{cut}}(A) := I(X_A ; X_{V \setminus A}) \tag{7}
Assuming the selected patch is locally Markovian at a time slice, the boundary shell \partial_R A strictly screens the interior A^\circ from the exterior V \setminus A, such that X_{A^\circ} \perp X_{V\setminus A} \mid X_{\partial_R A}. Consequently:
S_{\mathrm{cut}}(A) = I(X_{\partial_R A} ; X_{V \setminus A}) \le H(X_{\partial_R A}) \le |\partial_R A| \log q \tag{8}
Because Z_t is a capacity-limited compression of X_A, the data processing inequality guarantees I(Z_t ; X_{V \setminus A}) \le |\partial_R A| \log q. If the substrate graph G approximates a d-dimensional lattice, then |\partial_R A| \sim \operatorname{area}(A), not volume.
Thus, OPT rigorously yields a genuine Classical Boundary Law [39]. We can construct a formal epistemic ladder for future structural upgrades: 1. Classical Area Law: S_{\mathrm{cut}} \sim |\partial_R A| derived purely from locality and Markov screening. 2. Quantum Upgrade: Von Neumann entanglement entropy scaling becomes accessible only if the coarse predictive variables Z_t admit a formal Hilbert-space/Quantum Error Correction embedding. 3. Holographic Upgrade: True geometric holographic duality emerges only if we replace the bottleneck code Z_t with a hierarchical tensor network, reinterpreting S_{\mathrm{cut}} as a geometric min-cut.
By securing the classical boundary law first, OPT provides a strong mathematical floor—conditional on the Markov-screening assumption (X_{A^\circ} \perp X_{V \setminus A} \mid X_{\partial_R A})—from which the more speculative quantum formalisms can be safely constructed.
The action of the observer is formalized via the variational free energy F[q, \theta]:
F[q,\theta] = \mathbb{E}_q[-\log p_\theta(y_{1:T}, z_{1:T} \mid a_{1:T})] + \mathbb{E}_q[\log q(z_{1:T})] \tag{9}
Crucially, this enforces a strict mathematical separation: the substrate prior selects the hypothesis space, the virtual Stability Filter (4) bounds capacity-compatible structure, and FEP (9) governs agent-level inference inside that bounded structure. Physics emerges not as the Free Energy functional, but as the stable structure K_\theta that the Free Energy functional is successfully tracking.
Furthermore, sustaining this conscious render incurs an unavoidable thermodynamic cost. By Landauer’s Principle [52], each logically irreversible bit erasure dissipates at least k_B T \ln 2 of heat. Identifying one irreversible erasure per bottleneck update (a best-case bookkeeping assumption), the physical footprint of consciousness requires a minimum dissipation:
P_{\text{render}} \ge \dot{N}_{\text{erase}} \cdot k_B T \ln 2 \ge C_{\max} \cdot k_B T \ln 2 \tag{10}
This is a best-case lower bound under one-erase-per-update bookkeeping — not a generic consequence of bandwidth alone. The resulting bound (\sim 10^{-19} W) is vastly exceeded by actual neural dissipation (~20W), reflecting the enormous thermodynamic overhead of biological implementation. Equation (10) establishes the strict theoretical floor on the minimum possible physical footprint of any substrate instantiating a C_{\max}-bounded conscious render.
(Remark: The preceding thermodynamic and informational bounds strictly govern the real-time update bandwidth C_{\max}. However, this does not capture the full experiential dimensionality of the observer’s standing state, nor how the codec manages its own complexity over deep time. These structural mechanics—the Phenomenal State Tensor formulation of rich experience and the active maintenance cycle of sleep/dreaming—are fully derived in §3.5 and §3.6 below.)
3.5 The Phenomenal State Tensor and the Prediction Asymmetry
3.5.1 The Experiential Density Puzzle
The formal apparatus of §§3.1–3.4 successfully constrains the
update throughput of a conscious observer via the capacity
ceiling C_{\max} \approx
\mathcal{O}(10) bits/s.
However, phenomenal experience presents an immediate structural puzzle:
the felt richness of a single visual moment — the simultaneous presence
of colour, depth, texture, sound, proprioception, and affect — vastly
exceeds the information content that C_{\max} could deliver in any single update
window \Delta t \approx 50\
\text{ms}.
The maximum new information resolved per conscious moment is:
B_{\max} = C_{\max} \cdot \Delta t \approx 10\ \text{bits/s} \times 0.05\ \text{s} = 0.5\ \text{bits} \tag{T8-1}
This is far less than one bit of genuinely novel information per perceptual frame, yet the phenomenal scene appears informationally dense. To resolve this discrepancy without inflating the narrow update bandwidth, we must explicitly distinguish two structurally distinct quantities: 1. C_{\max} — the update throughput: the rate of prediction-error signal resolved into the settled causal record per unit time. 2. C_{\text{state}} — the standing-state complexity: the Kolmogorov complexity K(P_\theta(t)) of the generative model currently loaded and active.
These are not the same quantity. C_{\max} governs the gate; C_{\text{state}} characterises the room. The remainder of this section makes the distinction precise and introduces the Phenomenal State Tensor P_\theta(t) as the formal object corresponding to the standing inner scene.
3.5.2 The Prediction Asymmetry: Upward Errors and Downward Predictions
OPT inherits the predictive-processing architecture in which the codec K_\theta operates as a hierarchical generative model. Under this architecture, two distinct information flows traverse the Markov Blanket \partial_R A simultaneously:
Upward flow (prediction error, \varepsilon_t): the mismatch between K_\theta’s current prediction and the sensory signal arriving at \partial_R A. This is the correction signal. It is sparse, surprise-driven, and strictly capacity-limited.
Downward flow (prediction, \pi_t): the generative model’s active rendering of expected sensory states, propagated from higher to lower hierarchical levels. This is the scene itself. It is dense, continuous, and drawn from the full parameterisation of K_\theta.
Formally, let the sensory boundary state be X_{\partial_R A}(t), and let the codec’s predicted boundary state be:
\pi_t := \mathbb{E}_{K_\theta}\!\left[X_{\partial_R A}(t) \mid Z_t\right] \tag{T8-2}
The prediction error is then:
\varepsilon_t := X_{\partial_R A}(t) - \pi_t \tag{T8-3}
C_{\max} bounds the error signal, not the prediction. The mutual information between the error signal and the bottleneck state obeys:
I(\varepsilon_t\,;\,Z_t) \leq C_{\max} \cdot \Delta t = B_{\max} \tag{T8-4}
The prediction \pi_t, by contrast, is drawn from the full generative model and carries no such constraint. Its informational content is bounded only by the complexity of K_\theta itself. This asymmetry is the formal basis for distinguishing phenomenal richness from update bandwidth.
3.5.3 Definition: The Phenomenal State Tensor P_\theta(t)
We define the Phenomenal State Tensor P_\theta(t) natively as the full standing active parameter subset of the generative model deployed to project through the Markov Blanket at time t:
P_\theta(t) := \bigl\{\, K_\theta(\cdot,\, \cdot) \,\bigr\}_{\text{active}} \tag{T8-5}
That is, P_\theta(t) is the complete parameterized architecture the codec currently holds ready to generate predictions over the observable boundary states X_{\partial_R A}, evaluated independently of any single specific instantiation of the compressed latent state Z_t and action a_t. Its structural complexity is characterised naturally by the Kolmogorov complexity of this current standing parameter configuration:
C_{\text{state}}(t) := K\!\left(P_\theta(t)\right) \tag{T8-6}
where K(\cdot) denotes prefix Kolmogorov complexity. C_{\text{state}}(t) is the standing-state complexity — the number of bits of compressed structure the codec is currently holding in active deployment.
Upper bound on boundary channel flow. The mutual information between the bottleneck state and the boundary is bounded by standard Shannon inequalities [16] (Eq. 8 of the base paper):
I\!\left(Z_t\,;\,X_{\partial_R A}\right) \leq H\!\left(X_{\partial_R A}\right) \leq |\partial_R A|\cdot \log q \tag{T8-7}
This bounds the channel flow across the Markov Blanket — vastly large relative to B_{\max}. Important caveat: This is a bound on the Shannon-theoretic mutual information I(Z_t\,;\,X_{\partial_R A}), not a bound on the Kolmogorov complexity K(P_\theta(t)) of the standing model. Shannon entropy quantifies ensemble-average uncertainty; Kolmogorov complexity quantifies the description length of a specific computable object. No general inequality bridges these quantities without additional assumptions (e.g., a universal prior over model classes). We therefore do not claim that C_{\text{state}} \leq H(X_{\partial_R A}). The standing-state complexity C_{\text{state}} is bounded empirically (§3.10), not by the boundary entropy.
Heuristic lower bound on C_{\text{state}}. The Stability Filter directly constrains only the update rate R_{\text{req}} \leq B_{\max}, not the standing-model depth. However, a codec with insufficient structural complexity cannot generate accurate predictions \pi_t matching the statistics of a complex environment across the forward fan \mathcal{F}_h(z_t). This imposes a practical minimum on C_{\text{state}}: below some threshold, R_{\text{req}} would systematically exceed B_{\max} because the prediction errors \varepsilon_t would be persistently large. This lower bound is empirically motivated rather than formally derived — no closed-form expression C_{\text{state}} \geq f(R_{\text{req}}, \text{environment statistics}) is presently available.
3.5.4 Block’s Distinction as a Structural Corollary
The formal distinction between P_\theta(t) and Z_t maps precisely onto Ned Block’s distinction between phenomenal consciousness (P-consciousness) and access consciousness (A-consciousness) [47]:
| Block’s Category | OPT Object | Information Content | Bandwidth-Limited? |
|---|---|---|---|
| P-consciousness (qualia, felt scene) | P_\theta(t) | C_{\text{state}} = K(P_\theta(t)) \gg B_{\max} | No |
| A-consciousness (reportable content) | Z_t | B_{\max} = C_{\max} \cdot \Delta t \approx 0.5\ \text{bits} | Yes |
Under OPT, P-consciousness is the downward prediction \pi_t drawn from the full tensor P_\theta(t). A-consciousness is the bottleneck output Z_t — the thin slice of the scene that has been compressed sufficiently to enter the causal record \mathcal{R}_t and become available for report. The felt richness of a visual moment is P_\theta(t); the ability to say “I see red” requires that feature to pass through Z_t.
This corollary resolves the apparent paradox of a rich phenomenal scene sustained by a sub-bit update channel: the scene is not delivered through the channel each frame — it is already loaded in P_\theta(t). The channel updates it, incrementally and selectively, frame by frame.
3.5.5 The Update Dynamics of P_\theta(t)
The update rule for P_\theta(t) is governed by the prediction-error signal \varepsilon_t filtered through the bottleneck:
P_\theta(t+1) = \mathcal{U}\!\left(P_\theta(t),\, \varepsilon_t,\, Z_t\right) \tag{T8-8}
where \mathcal{U} is the codec’s learning operator — in Active Inference terms, the gradient step on variational free energy \mathcal{F}[q, \theta] (Eq. 9 of base paper) restricted by the capacity constraint I(X_t\,;\,Z_t) \leq B.
The key structural property is that \mathcal{U} is selective: only those regions of P_\theta(t) implicated by the current prediction error \varepsilon_t are updated. The remainder of the standing tensor is held constant across the frame. This gives the conscious moment its characteristic structure: a stable phenomenal background against which a small foreground of resolved novelty is laid.
The codec thus implements a form of sparse update on a dense prior — a design principle that maximises phenomenal coherence per unit of update bandwidth.
3.5.6 Scope and Epistemic Status
The Phenomenal State Tensor P_\theta(t) is a formal characterisation of the structural shadow the phenomenal scene must cast, consistent with the Agency Axiom (§3.6). It does not resolve the Hard Problem. OPT continues to treat phenomenal consciousness as an irreducible primitive; P_\theta(t) specifies the geometry of the container, not the nature of its contents.
The claim is structural and falsifiable in the following sense: if the qualitative richness of reported experience (as operationalised through, e.g., measures of phenomenal complexity in psychophysical tasks) correlates with codec depth — the hierarchical complexity of K_\theta as measurable via neural markers of predictive hierarchy — rather than with update bandwidth C_{\max}, then the P_\theta\,/\,Z_t distinction is empirically supported. Psychedelic states, which dramatically alter the structure of K_\theta without consistently altering behavioural throughput, represent a natural test domain.
3.6 The Codec Lifecycle: The Maintenance Cycle Operator \mathcal{M}_\tau
3.6.1 The Static Codec Problem
The framework of §§3.1–3.5 treats K_\theta and its realisation P_\theta(t) as dynamic across update frames but implicitly assumes the codec’s structural architecture — the parameter space \Theta itself — is fixed. This is adequate for a synchronic analysis of a single conscious moment, but inadequate for a theory of consciousness across deep time.
A codec operating continuously accumulates structural complexity: every learned pattern adds parameters to K_\theta, increasing C_{\text{state}}(t). Without a mechanism for controlled complexity reduction, C_{\text{state}} would grow monotonically until the codec exceeded its thermodynamic runability ceiling — the point at which the metabolic cost of maintaining P_\theta(t) exceeds the organism’s energy budget, or the internal complexity of K_\theta exceeds the Stability Filter’s capacity-compatible description length.
This section introduces the Maintenance Cycle Operator \mathcal{M}_\tau — the formal mechanism by which the codec manages its own complexity across time, operating primarily during states of reduced sensory load (paradigmatically: sleep).
3.6.2 The Maintenance Condition
Define the codec runability condition as the requirement that the Kolmogorov complexity of the current generative model remain below a structural ceiling C_{\text{ceil}} set by the organism’s thermodynamic budget:
K\!\left(P_\theta(t)\right) \leq C_{\text{ceil}} \tag{T9-1}
C_{\text{ceil}} is not the same as C_{\max}. It is a much larger quantity — the total structural complexity the codec can sustain in its parameter space — but it is finite. Violations of (T9-1) correspond to cognitive overload, memory interference, and ultimately to the pathological case described by Borges’ [53] Funes the Memorious: a system that has acquired so much uncompressed detail that it can no longer function predictively.
The Maintenance Cycle Operator \mathcal{M}_\tau is defined as acting during periods when R_{\text{req}} \ll C_{\max} — specifically, when the required predictive rate drops sufficiently that the bandwidth freed can be redirected to internal restructuring:
\mathcal{M}_\tau : P_\theta(t) \;\longrightarrow\; P_\theta(t + \tau) \qquad \text{during} \quad R_{\text{req}}(t) \ll C_{\max} \tag{T9-2}
\mathcal{M}_\tau decomposes into three structurally distinct passes, each targeting a different aspect of codec complexity management.
3.6.3 Pass I — Pruning (Forgetting as Active MDL Pressure)
The first pass applies Minimum Description Length (MDL) pressure to the current codec parameters. For each component \theta_i of the generative model K_\theta, define its predictive contribution as the mutual information it provides about the future observation stream, net of the storage cost of retaining it:
\Delta_{\mathrm{MDL}}(\theta_i) := I\!\left(\theta_i\,;\,X_{t+1:t+\tau} \mid \theta_{-i}\right) - \lambda \cdot K(\theta_i) \tag{T9-3}
where \theta_{-i} denotes all parameters except \theta_i, \lambda is a retention threshold (bits of future prediction bought per bit of model complexity), and K(\theta_i) is the description length of the component.
The pruning rule is:
\text{Prune } \theta_i \quad \text{if} \quad \Delta_{\mathrm{MDL}}(\theta_i) < 0 \tag{T9-4}
That is, discard \theta_i when its predictive contribution per bit of storage falls below the threshold \lambda. This is forgetting formalised not as failure but as thermodynamically rational erasure: each pruned component recovers K(\theta_i) bits of model capacity for reuse.
By Landauer’s Principle [52], each pruning operation establishes a thermodynamic floor for erasure:
W_{\text{prune}}(\theta_i) \geq K(\theta_i) \cdot k_B T \ln 2 \tag{T9-5}
While actual biological metabolism operates many orders of magnitude above this theoretical minimum (Watts versus femtowatts) due to severe implementation overhead, the structural necessity of the cost remains. Sleep therefore carries a fundamental thermodynamic signature in OPT: it is a period of net information erasure whose energy cost is mandated by physics rather than merely biological inefficiency.
The aggregate complexity reduction of the pruning pass is:
\Delta K_{\text{prune}} = \sum_i K(\theta_i)\cdot \mathbf{1}\!\left[\Delta_{\mathrm{MDL}}(\theta_i) < 0\right] \tag{T9-6}
3.6.4 Pass II — Consolidation (Learning as Compression Gain)
The pruning pass removes components with insufficient predictive return. The consolidation pass reorganises the remaining components into more compressed representations.
During waking operation, the codec acquires patterns under real-time pressure: each update must be computed within \Delta t, leaving no time for global structural reorganisation of K_\theta. Recently acquired patterns are stored in a relatively uncompressed form — high K(\theta_{\text{new}}) for the predictive contribution they provide. The consolidation pass applies offline MDL compression to these recent acquisitions.
Let \Theta_{\text{recent}} \subset \Theta denote the set of parameters acquired since the last maintenance cycle. The consolidation operator finds the minimum-complexity reparameterisation \theta' of \Theta_{\text{recent}} such that the predictive distribution it generates is within tolerable distortion D_c of the original:
\theta'_{\text{cons}} = \arg\min_{\theta'} K(\theta') \quad \text{s.t.} \quad D_{\mathrm{KL}}\!\left(P_{\theta'}(\cdot) \,\Big\|\, P_{\Theta_{\text{recent}}}(\cdot)\right) \leq D_c \tag{T9-7}
The compression gain recovered is:
\Delta K_{\text{compress}} = K(\Theta_{\text{recent}}) - K(\theta'_{\text{cons}}) \tag{T9-8}
\Delta K_{\text{compress}} is the number of bits of model capacity recovered by reorganising recent experience into more efficient representations. Each unit of \Delta K_{\text{compress}} directly reduces the future R_{\text{req}} for similar environments — the codec becomes cheaper to run in familiar territory.
This formalises the empirically observed function of hippocampal-neocortical memory consolidation during slow-wave sleep: the transfer from high-bandwidth episodic storage (hippocampus, high K) to compressed semantic storage (neocortex, low K) is precisely the compression operation of (T9-7). The prediction is that compression gain \Delta K_{\text{compress}} should correlate with the degree of behavioural improvement observed after sleep on tasks involving structured pattern recognition.
3.6.5 Pass III — Forward Fan Sampling (Dreaming as Adversarial Self-Testing)
The third pass operates primarily during REM sleep, when sensory input is actively gated and motor output is inhibited. Under these conditions, R_{\text{req}} \approx 0: the codec is receiving no correction signal from the external environment. The full bandwidth budget C_{\max} is available for internal operation.
OPT frames this state formally as unconstrained forward-fan exploration: the codec generates trajectories through \mathcal{F}_h(z_t) — the set of admissible future sequences (Eq. 5 of base paper) — without anchoring those trajectories to real incoming data. This is simulation: the codec runs its generative model K_\theta forward in time, unimpeded by reality.
The sampling distribution over the fan is not uniform. Define the importance weight of a branch b \in \mathcal{F}_h(z_t) as:
w(b) := \exp\!\left(\beta\cdot |E(b)|\right) \tag{T9-9}
where \beta is an inverse temperature parameter and E(b) is the emotional valence of the branch, defined as:
E(b) := -\log P_{K_\theta}(b \mid z_t) + \alpha \cdot \mathrm{threat}(b) \tag{T9-10}
The first term -\log P_{K_\theta}(b \mid z_t) is the negative log-probability of the branch under the current codec — its surprise value. The second term \mathrm{threat}(b) is a fitness-relevant consequence measure formally defined as the expected increase in required predictive rate if the codec were to traverse branch b:
\mathrm{threat}(b) := \mathbb{E}\!\left[\, R_{\text{req}}(D_{\min} \mid b) - R_{\text{req}}(D_{\min} \mid z_t)\,\right] \tag{T9-10a}
That is, \mathrm{threat}(b) quantifies the degree to which branch b, if realised in waking life, would push the codec toward or beyond its bandwidth ceiling B_{\max} — through physical harm, social rupture, or narrative collapse that would force costly model revision. Branches with \mathrm{threat}(b) > B_{\max} - R_{\text{req}}(D_{\min} \mid z_t) are existentially threatening: they would violate the Stability Filter condition. The weighting parameter \alpha \geq 0 controls the relative influence of consequence versus surprise in the sampling distribution.
The sampling operator draws branches proportional to w(b):
b_{\text{sample}} \sim \mathcal{F}_h(z_t) \quad \text{with probability} \propto w(b) \tag{T9-11}
This implements importance-weighted forward-fan sampling: the codec disproportionately rehearses branches that are either highly surprising or highly consequential, regardless of their base-rate probability. Low-probability, high-threat branches — precisely those for which the codec is least prepared — receive the greatest sampling attention.
Each sampled branch is then evaluated for coherence under K_\theta. Branches that generate incoherent prediction sequences — where the codec’s own generative model cannot maintain narrative stability — are identified as brittleness points: regions of the forward fan where the codec would fail if the branch were encountered in waking life. The codec can then update P_\theta to reduce K_\theta’s vulnerability at those points, before being exposed to them with real thermodynamic stakes.
Dreaming is therefore adversarial self-testing of the codec at zero risk. The functional consequence is a codec that is systematically better prepared for the low-probability, high-consequence branches of its own forward fan. This OPT framing provides an information-theoretic grounding for Revonsuo’s [46] threat-simulation theory of dreaming, extending it from an evolutionary-functional account to a formal structural necessity: any codec operating under the Stability Filter must periodically stress-test its own forward fan, and the offline maintenance state is the only period when this can be done without real-world thermodynamic cost.
Emotional tagging as a retention weight prior. In the waking state, the emotional valence E(b) computed during REM sampling serves as a prior retention weight biasing the MDL threshold \lambda in (T9-3). Experiences with high |E(b)| — strongly surprising or consequential — are assigned a higher effective \lambda, making them more resistant to pruning in the next maintenance cycle. This is the formal account of emotional memory enhancement: affect is not noise contaminating the memory system; it is the codec’s relevance signal, marking patterns whose predictive value exceeds their base-rate statistical frequency.
3.6.6 The Full Maintenance Cycle and Net Complexity Budget
The three passes of \mathcal{M}_\tau compose sequentially. The net effect on codec complexity across one maintenance cycle of duration \tau is:
K\!\left(P_\theta(t+\tau)\right) = K\!\left(P_\theta(t)\right) - \Delta K_{\text{prune}} - \Delta K_{\text{compress}} + \Delta K_{\text{REM}} \tag{T9-12}
where \Delta K_{\text{REM}} is the small positive increment from patterns newly consolidated from the REM sampling pass — those brittleness-point repairs that required new parameter updates.
For a stable cognitive system operating across years, the long-run budget requires:
\left\langle \Delta K_{\text{prune}} + \Delta K_{\text{compress}} \right\rangle \geq \left\langle \Delta K_{\text{waking}} + \Delta K_{\text{REM}} \right\rangle \tag{T9-13}
where \Delta K_{\text{waking}} is the complexity acquired during the preceding waking period. Inequality (T9-13) is the formal statement that maintenance must keep pace with acquisition. Chronic sleep deprivation, in OPT terms, is not merely fatigue — it is progressive complexity overflow: the codec approaches C_{\text{ceil}} while its pruning and consolidation budget is insufficient to restore headroom.
3.6.7 Empirical Predictions
The Maintenance Cycle framework generates the following testable structural expectations:
Sleep duration scales with codec complexity. Organisms or individuals who acquire more structured information during waking periods should require proportionally longer or deeper maintenance cycles. The prediction is not simply that hard cognitive work requires more sleep (which is established), but that the type of learning matters: pattern-rich, compressible learning should require less consolidation time than unstructured, high-entropy experience, because \Delta K_{\text{compress}} is larger in the former case.
REM content is importance-weighted over the forward fan, not frequency-weighted. Dream content should disproportionately sample low-probability, high-consequence branches relative to their waking frequency. This is consistent with the empirical predominance of threat, social conflict, and novel-environment content in dream reports — the codec samples what it needs to stress-test, not what it most often encounters.
Compression efficiency improves post-sleep proportional to \Delta K_{\text{compress}}. The specific prediction is that post-sleep performance improvements should be largest on tasks requiring structural generalisation (i.e., applying a compressed rule to new instances) rather than simple repetition — because \Delta K_{\text{compress}} specifically reorganises \Theta_{\text{recent}} into more generalisable forms.
Pathological rumination corresponds to REM sampling stuck at high-|E| branches. If the importance-weighting parameter \beta is pathologically elevated, the sampling distribution over \mathcal{F}_h(z_t) concentrates on high-threat branches to the exclusion of repair. The codec spends its maintenance cycle repeatedly sampling the same threatening branches without successfully reducing their surprise value — the formal structure of anxiety and PTSD nightmares.
3.6.8 Relationship to the Phenomenal State Tensor
\mathcal{M}_\tau acts on P_\theta(t) as defined in §3.5: it restructures the standing-state complexity C_{\text{state}} across the maintenance window. The temporal profile of P_\theta(t) under \mathcal{M}_\tau is:
- Waking acquisition: C_{\text{state}} increases at rate bounded by the learning operator \mathcal{U} (Eq. T8-8), as new patterns are incorporated into K_\theta.
- Slow-wave sleep (Passes I–II): C_{\text{state}} decreases as pruning and consolidation recover model capacity.
- REM (Pass III): C_{\text{state}} undergoes selective local increase at brittleness points, with net effect small relative to the reductions of Passes I–II.
The conscious experience corresponding to each phase is consistent with this structure: waking life accumulates the richness of P_\theta(t); slow-wave sleep is phenomenally sparse or absent (consistent with minimal P_\theta(t) activation during structural reorganisation); REM presents a phenomenally vivid but internally generated scene (Pass III running the full generative model forward in the absence of sensory correction).
Summary: New Formal Objects Introduced
| Symbol | Name | Definition | Equation |
|---|---|---|---|
| P_\theta(t) | Phenomenal State Tensor | Full activation of K_\theta at time t, projected through \partial_R A | T8-5 |
| C_{\text{state}}(t) | Standing-state complexity | K(P_\theta(t)), Kolmogorov complexity of active codec | T8-6 |
| \pi_t | Downward prediction | \mathbb{E}_{K_\theta}[X_{\partial_R A}(t) \mid Z_t], the rendered scene | T8-2 |
| \varepsilon_t | Prediction error (upward) | X_{\partial_R A}(t) - \pi_t, novelty signal bounded by C_{\max} | T8-3 |
| \mathcal{M}_\tau | Maintenance Cycle Operator | P_\theta(t) \to P_\theta(t+\tau) under low R_{\text{req}} | T9-2 |
| \Delta_{\mathrm{MDL}}(\theta_i) | MDL retention score | Predictive contribution minus storage cost | T9-3 |
| E(b) | Branch emotional valence | Surprise plus weighted threat of branch b | T9-10 |
| w(b) | Branch importance weight | \exp(\beta \cdot |E(b)|), drives REM sampling distribution | T9-9 |
| \Delta K_{\text{prune}} | Pruning complexity recovery | Bits recovered by forgetting below-threshold components | T9-6 |
| \Delta K_{\text{compress}} | Consolidation compression gain | Bits recovered by MDL recompression of recent acquisitions | T9-8 |
3.7 The Tensor-Network Mapping: Inducing Geometry from Code Distance
The Epistemic Ladder introduced in §3.4 establishes a rigorous Classical Boundary Law (S_{\mathrm{cut}} \sim |\partial_R A|). However, to fully bridge the Ordered Patch Theory strictly to the geometrization of quantum information (e.g., AdS/CFT and the Ryu-Takayanagi formula), we must formally upgrade the structure of the latent code Z_t.
If we formally postulate that the bottleneck mapping q^\star(z \mid X_t) does not simply extract a flat list of features, but operates via a recursive, coarse-graining renormalization group flow, the generative model structurally aligns onto the geometry of a hierarchical tensor network \mathcal{T} (akin to MERA [43] or HaPY networks [44]). (Remark: Appendix T-3 formally derives a structural homomorphic correspondence between the Stability Filter’s coarse-graining cascade and the MERA network geometry bounding, strictly mapping the Informational Causal Cone to the equivalent MERA causal cone). The boundary states of this network are precisely the screened Markov boundary states X_{\partial_R A}. The network \mathcal{T} acts as a bulk geometry whose “depth” represents the layers of computational coarse-graining required to compress the boundary into the minimal bottleneck state Z_t.
Under this tensor-network upgrade, the predictive cut entropy S_{\mathrm{cut}}(A) across the boundary transforms mathematically into the minimum number of tensor bonds that must be severed to isolate the subregion A. Let \chi be the bond dimension of the network. The capacity bound internally maps as:
S_{\mathrm{cut}}(A) \le |\gamma_A| \log \chi \tag{11}
where \gamma_A is the minimal-cut surface through the inner deep layer bulk data structure of \mathcal{T}. This is explicitly a discrete structural analogue of the bulk minimal-cut layer mapped by the Ryu-Takayanagi holographic entropy bound. Appendix P-2 (Theorem P-2d) formally establishes the full discrete quantum RT formula S_{\text{vN}}(\rho_A) \leq |\gamma_A| \log \chi via the Schmidt rank of the MERA state, conditional on the local noise model and QECC embedding derived therein. The continuum limit upgrading this to the full Ryu-Takayanagi formula with bulk correction term remains an open edge.
Crucially, in OPT, this “bulk space” is not a pre-existing physical container. It is the strictly informational metric space of the observer’s codec. The emergent phenomenological spacetime geometry “curves” precisely where the required code distance diverges to resolve overlapping internal causal states. This Tensor-Network formalism illustrates a formal path by which OPT might induce spatial geometry directly from the error-correction distances intrinsically mandated by the Stability Filter, offering a constructive conjecture that holographic spacetime models optimal data-compression formats.
3.8 The Agency Axiom & The Phenomenal Residual
The mathematical apparatus developed in Sections 3.1–3.7 precisely defines the geometry of the observer’s reality—the tensor network, the predictive cut, and the causal cone. However, what is the nature of the primitive interiority that experiences the passage through it? We formally define this via the Agency Axiom: the traversal of the C_{\max} aperture is intrinsically a phenomenological event.
While we take the presence of subjective feeling as axiomatic, Theorem P-4 (The Phenomenal Residual) identifies its rigorous structural correlate. Because the bounded codec actively perturbs the boundary \partial_R A, stable prediction within C_{\max} limits requires it to model the consequences of its own future actions. Thus, the codec K_{\theta} must maintain a predictive self-model \hat{K}_{\theta}. However, by the algorithmic bounds of informational containment [13], a finite computational system cannot contain a complete structural representation of itself; the internal model is rigidly bounded to a lower complexity than the parent codec (K(\hat{K}_{\theta}) < K(K_{\theta})).
This necessitates an irreducible Phenomenal Residual (\Delta_{\text{self}} > 0). This un-modellable residual acts as the computational “blind spot” within the active inference cycle. Because it exists in the informational shadow exceeding the computational reach of the self-model, it is inherently ineffable; because it exists as the localized delta between a specific codec and its model, it is computationally private; and dictated by fundamental limits on self-reference and necessary variational approximation, it is non-eliminable. The topological narrowing at the C_{\max} aperture is intrinsically correlated with the mathematical necessity of an incomplete algorithm undergoing its own boundaries. The math describes the formal contour of the experience, and the Agency Axiom asserts that this residual locus constitutes the subjective “I”. (See Appendix P-4 for the formal derivation).
The Informational Maintenance Circuit
Within a single update frame [t, t+\Delta t], the observer executes the following closed causal circuit:
P_\theta(t) \;\xrightarrow{\ \pi_t\ }\; \partial_R A \;\xrightarrow{\ \varepsilon_t\ }\; Z_t \;\xrightarrow{\ \mathcal{U}\ }\; P_\theta(t+1) \tag{T6-1}
Explicitly:
Prediction (downward): The current tensor P_\theta(t) generates the predicted boundary state \pi_t = \mathbb{E}_{K_\theta}[X_{\partial_R A}(t) \mid Z_t] — the rendered scene.
Error (upward): The actual boundary state X_{\partial_R A}(t) arrives; the prediction error \varepsilon_t = X_{\partial_R A}(t) - \pi_t is computed.
Compression: \varepsilon_t is passed through the bottleneck to yield Z_t, the capacity-limited update token, with I(\varepsilon_t\,;\,Z_t) \leq B_{\max}.
Update: The learning operator \mathcal{U}(P_\theta(t), \varepsilon_t, Z_t) revises P_\theta(t+1), selectively modifying only those regions of the tensor implicated by \varepsilon_t.
Action: Simultaneously, P_\theta(t) selects action a_t via active inference descent on the variational free energy \mathcal{F}[q,\theta] (Eq. 9 of base paper), which alters the sensory boundary at t+1, influencing the next \varepsilon_{t+1}.
Interpretive note on the action step. The language of step 5 — “selects action” and “alters the sensory boundary” — is inherited from the Free Energy Principle’s standard active inference formalism, which assumes a physical environment that the agent pushes against via active states. Under OPT’s native render ontology (§8.6), a deeper reading applies: there is no independent external world against which the codec exerts force. What is experienced as “action” is a branch selection within the Forward Fan \mathcal{F}_h(z_t); the physical consequences of that selection arrive as subsequent input \varepsilon_{t+1}. The Markov blanket \partial_R A is not a two-way physical interface but the surface across which the selected branch delivers its next segment. This interpretive shift changes nothing in the mathematics of (T6-1)–(T6-3); it clarifies the ontological status of the action step within OPT’s framework. The mechanism of branch selection itself is addressed below.
This is the within-frame informational maintenance circuit: a closed causal mechanism in which the system’s internal model calculates localized structural predictions bounding boundary gradients, reads the error, and selectively updates itself. The loop is strictly informational and self-referential in the formal sense: P_\theta(t) determines both the structural prediction \pi_t and, via action a_t, a predictive component of the next sequential data stream input X_{\partial_R A}(t+1). (Note explicitly: this purely statistical screening layer is defined rigorously by informational Markov boundaries decoupling dynamics cleanly, differing inherently from complex biological autopoiesis where cell structures mechanically manufacture their own organic mass networks).
The Structural Viability Condition
The circuit (T6-1) is structurally viable if and only if it can sustain itself without the codec’s informational complexity exceeding its local runability limits. Formally:
K\!\left(P_\theta(t)\right) \leq C_{\text{ceil}} \quad \forall\, t \tag{T6-2}
where C_{\text{ceil}} is a heuristic parameter bounding the maximum structural complexity the codec can sustain. In principle, C_{\text{ceil}} should be derivable from the organism’s thermodynamic budget via Landauer’s principle (see the sketch in §3.10), but the full derivation chain — from metabolic power to erasure cost to maximum sustainable program complexity — is not yet formalised within OPT. C_{\text{ceil}} therefore remains an empirically motivated but formally underdetermined bound. A system satisfying (T6-2) operates as a structurally closed observer in OPT’s formal sense.
When (T6-2) is violated — when K(P_\theta(t)) \to C_{\text{ceil}} — the codec cannot maintain stable predictions across \mathcal{F}_h(z_t), R_{\text{req}} begins to exceed B_{\max}, and the Stability Filter condition fails. Narrative coherence collapses: the observer exits the set of observer-compatible streams.
The Maintenance Cycle \mathcal{M}_\tau (§3.6) is the mechanism that enforces (T6-2) over deep time, keeping K(P_\theta) within bounds via pruning, consolidation, and forward-fan stress-testing. Within-frame, (T6-2) is maintained by the selectivity of \mathcal{U}: the update operator modifies only the regions of P_\theta(t) implicated by \varepsilon_t, avoiding gratuitous complexity growth per frame.
Agency as Constrained Free Energy Minimisation
Within this structure, agency can be given a precise formal definition that is compatible with — but not reductive of — the Agency Axiom.
At the systems level, agency is the selection of action sequence \{a_t\} that minimises expected variational free energy subject to the informational viability condition:
a_t^\star = \arg\min_{a_t} \;\mathbb{E}\!\left[\mathcal{F}[q, \theta]\right] \quad \text{subject to} \quad K\!\left(P_\theta(t)\right) \leq C_{\text{ceil}} \tag{T6-3}
This is constrained active inference: the observer navigates the forward fan \mathcal{F}_h(z_t) not merely to minimise prediction error, but to minimise prediction error while keeping the codec viable. Branches that would temporarily reduce \varepsilon but drive K(P_\theta) toward C_{\text{ceil}} are penalised by the constraint. The observer preferentially selects branches along which it can continue to exist as a coherent observer.
This is the formal content of the intuition that agency is self-preserving navigation: the codec selects the branches of the forward fan along which it can continue to compress the world.
At the phenomenological level, the Agency Axiom remains untouched: phenomenal consciousness is the irreducible interiority of aperture-traversal; (T6-3) describes the structural shadow that traversal casts, not its inner nature.
Branch Selection as \Delta_{\text{self}} Execution
The constrained active inference formula (T6-3) specifies the objective of branch selection: minimise expected free energy subject to viability. The self-model \hat{K}_\theta evaluates branches of the Forward Fan by simulating their consequences. But Theorem P-4 establishes that K(\hat{K}_\theta) < K(K_\theta) — the self-model is necessarily incomplete. This incompleteness has a direct consequence for the branch selection problem: the self-model constrains the region from which selection can be drawn, but cannot fully specify the selection itself.
The actual moment of branch selection — the transition from the evaluated menu to the singular trajectory that enters the causal record — occurs in \Delta_{\text{self}}, the informational residual between the codec and its self-model. This is not a gap in the formalism; it is a structural necessity. Any attempt to fully specify the selection mechanism from within would require K(\hat{K}_\theta) = K(K_\theta), which P-4 proves is impossible for any finite self-referential system.
This has three immediate consequences:
Will and consciousness share the same structural address. The Hard Problem (why does traversal feel like something?) and the branch selection problem (what selects?) both point to \Delta_{\text{self}}. They are not two mysteries but two aspects of the same structural feature — the unmodelable gap between what the codec is and what it can model about itself.
The irreducibility of agency is explained, not merely asserted. The phenomenological experience of will — the irreducible sense that I chose — is the first-person signature of a process executing in the observer’s own blind spot. Any theory claiming to fully specify the selection mechanism has either eliminated \Delta_{\text{self}} (making the system a fully self-transparent automaton, which P-4 forbids) or is describing the self-model’s evaluation of branches and mistaking it for the selection itself.
Creativity as expanded \Delta_{\text{self}}. Near-threshold operation (R_{\text{req}} \to C_{\max}) strains the self-model’s capacity, effectively expanding the region of \Delta_{\text{self}} from which selection is drawn. This produces branch selections that are less predictable from the self-model’s perspective — experienced as creative insight, spontaneity, or “flow.” Conversely, the hypnagogic state (§3.6.5) relaxes the self-model from below, achieving the same expansion by a complementary route.
The self as residual. The experienced self — the continuous narrative of “who I am,” with stable preferences, a history, and a projected future — is \hat{K}_\theta’s running model of K_\theta: a compressed approximation that is always behind the codec it models (by the temporal lag inherent in self-reference). But the actual locus of experience, selection, and identity is \Delta_{\text{self}}: the part of the codec the narrative cannot reach. The self you know is your model of yourself; the self that knows is the gap the model cannot cross. This is the formal content of the contemplative discovery — across traditions, independently — that the ordinary sense of self is constructed and that beneath it is something that cannot be found as an object (see Appendix T-13, Corollary T-13c).
Deliberation is real but incomplete. The self-model’s evaluation of the Forward Fan is a genuine computational process that shapes the outcome. Deliberation constrains the basin of attraction within which \Delta_{\text{self}} operates: a more developed codec narrows the viable branches that selection can land on. But the final transition — why this branch rather than that one, among the viable set — is structurally opaque to the deliberating self. This is why deliberation feels both causally efficacious and phenomenologically incomplete: the observer correctly senses that its reasoning matters, but also correctly senses that something beyond the reasoning finalises the choice.
The Strange Loop as Formal Closure
The self-referential structure of (T6-1) instantiates Hofstadter’s [45] Strange Loop in a precise information-theoretic form. The loop is strange in the following sense: P_\theta(t) contains, as a substructure, a model of the codec’s own future states — the forward-fan sampling of Pass III (\mathcal{M}_\tau, §3.6.5) is precisely the codec running a simulation of itself encountering future branches. The system models its own model.
The formal closure this provides: the informationally closed observer is not merely a system that maintains a boundary against external noise; it is a system whose boundary-maintenance is partly constituted by its model of what that boundary needs to be in the future. The strange loop is not an optional add-on to the framework; it is the structural mechanism by which the viability condition (T6-2) is enforced proactively rather than reactively. An observer that could not simulate its own future codec states could not prepare for the brittleness points identified in Pass III, and would be systematically more vulnerable to narrative collapse.
The structural requirements of (T6-1)–(T6-3) function as necessary preconditions for self-referential closure. While simple forward prediction (e.g., a chess engine’s look-ahead) constitutes planning rather than genuine self-reference, the OPT codec goes further: P_\theta(t) contains a sub-model whose output modifies the distributions governing its own future states \{P_\theta(t+h)\}_{h>0}. This structural self-modeling is functionally necessary for long-run stability — a codec unable to anticipate its own approaching viability limits cannot prepare for the brittleness points identified in Pass III (§3.6.5), and will systematically collapse against the (T6-2) ceiling in non-stationary environments.
Epistemic Scope: Formally Scoping Agency Reductionism
This formalisation precisely delineates what OPT achieves at the systems level: it identifies the structural conditions an observer must satisfy to maintain boundary viability. This Formally Scopes the Agency Reductionism Problem without claiming to resolve it.
The scoping is genuine, not definitional. The systems-level description (T6-1)–(T6-3) exhaustively characterises the structural shadow of agency — the information-theoretic constraints any boundary-maintaining observer must satisfy. The Agency Axiom occupies the complementary domain: phenomenal consciousness is the irreducible interiority of aperture-traversal, and the formalisation above describes only the shape of the container, not the nature of what it contains. The Hard Problem is thereby located at a precise structural locus (the C_{\max} aperture) rather than dissolved or declared solved.
3.9 Free Will and the Phenomenological Menu
The isolation of the traversal mechanism fundamentally clarifies the nature of agency. In the Active Inference loop (Equation 9), the observer must execute a policy sequence \{a_t\}. Under reductive physicalism, the selection of the action a_t is determined (or randomly sampled) by the underlying physics, rendering free will an illusion or a mere linguistic redefinition.
OPT reverses this dependency. Because the localized “physics” of the patch is merely the generative model’s predictive estimation of the substrate, the physical laws only constrain the Forward Fan \mathcal{F}_h(z_t) to a set of macroscopic probabilities. Crucially, unless the patch is a perfectly predictable automaton (which violates the thermodynamic requirement for generative structural complexity), the Forward Fan contains genuine, unresolved branch multiplicity from the observer’s limited perspective.
Since the descriptive physics merely outlines the menu of these valid branches, it cannot logically experience the selection. On the compatibilist reading developed further in §8.6, the branch path is mathematically fixed in the timeless substrate; selection is the phenomenological experience of traversal. From the third-person perspective (the outside geometry), branch-selection appears as spontaneous noise, quantum collapse, or statistical fluctuation. From the first-person internal perspective, the boundaries of uncertainty guarantee that the traversal is experienced as the exertion of Will—the primitive action of navigating the uncompressed frontier. In OPT, free will is not a contra-causal breach of physical law; it is the necessary phenomenological openness experienced by a bounded observer collapsing a formal menu into a singular rendered timeline.
The render-ontology sharpening. Under OPT’s native ontology (§8.6), the distinction between perception and action dissolves at the substrate level. What is experienced as “output” — reaching, deciding, choosing — is stream content that the codec is navigating. The codec does not act on the world; it traverses a branch of \mathcal{F}_h(z_t) in which the experience of acting is part of what arrives at the boundary. What the Free Energy Principle calls active states — the outward flow modifying the environment — are, in OPT’s render ontology, the codec’s branch selection expressing itself as subsequent input content. The Markov blanket is the surface across which the selected branch delivers its next segment, not a membrane through which the observer pushes against an external reality. This sharpens the compatibilist account: there is no distinction between perceived and willed at the substrate level; both are stream content; the phenomenological distinction arises from how P_\theta(t) tags certain content as “self-initiated” — a tagging whose mechanism, like all branch selection, ultimately executes in \Delta_{\text{self}} (§3.8).
3.10 The Informational Cost of the Render and the Three-Level Bound Gap
The defining mathematical boundary of the Ordered Patch Theory is the formal comparison of informational generating costs.
Let U_{\text{obj}} be the full informational state of an objective universe. The Kolmogorov complexity K(U_{\text{obj}}) is astronomically high. Let S_{\text{obs}} be the localized, low-bandwidth stream experienced by an observer (strictly bounded by the \mathcal{O}(10) bits/s threshold). In OPT, the universe U_{\text{obj}} does not exist as a rendered computational object. The apparent “objective universe” is instead the internal Generative Model constructed by Active Inference.
The Bekenstein Bound for a Biologically Realistic Observer
The Bekenstein bound [40] gives the maximum thermodynamic entropy — equivalently, the maximum information content — of any physical system bounded by radius R with total energy E:
S_{\text{Bek}} \leq \frac{2\pi R E}{\hbar c} \tag{T7-1}
For a human brain as the observer’s Markov Blanket boundary \partial_R A:
- Bounding radius: R \approx 0.07\ \text{m}
- Total rest-mass energy: E = m c^2 \approx 1.4\ \text{kg} \times (3 \times 10^8\ \text{m/s})^2 = 1.26 \times 10^{17}\ \text{J}
- Reduced Planck constant: \hbar = 1.055 \times 10^{-34}\ \text{J}\cdot\text{s}
- Speed of light: c = 3 \times 10^8\ \text{m/s}
Substituting:
S_{\text{Bek}} = \frac{2\pi \times 0.07 \times 1.26 \times 10^{17}}{1.055 \times 10^{-34} \times 3 \times 10^8} = \frac{5.54 \times 10^{16}}{3.17 \times 10^{-26}} \approx 1.75 \times 10^{42}\ \text{nats} \tag{T7-2}
Converting to bits (dividing by \ln 2):
S_{\text{Bek}} \approx 2.52 \times 10^{42}\ \text{bits} \tag{T7-3}
The holographic area bound, S \leq A / 4l_P^2, yields a larger figure. For a sphere of radius R = 0.07\ \text{m}, surface area A = 4\pi R^2 \approx 0.062\ \text{m}^2, and Planck length l_P = 1.616 \times 10^{-35}\ \text{m}:
S_{\text{holo}} = \frac{0.062}{4 \times (1.616 \times 10^{-35})^2} = \frac{0.062}{1.044 \times 10^{-69}} \approx 5.9 \times 10^{67}\ \text{bits} \tag{T7-4}
We adopt the formulation bounded by (T7-3) tracking explicitly S_{\text{phys}} \approx 2.5 \times 10^{42}\ \text{bits} for the structural framework of this analysis. We explicitly flag structurally that using the total rest-mass energy E=mc^2 inflates this metric to an extreme maximal upper limit; active internal biological thermodynamic interactions utilizing purely internal chemical energy bounds (\sim 10-100\text{J}) drop this Bekenstein limit dramatically closer to \sim 10^{26} bits. The qualitative structural gap mechanism formally demonstrated below holds equivalently utilizing any parameter formulation of these physical upper bounds across all margins, acting formally as a conservative limit holding a fortiori against extreme pure geometric Holographic equivalents mapped previously (T7-4).
The Three-Level Gap
The Phenomenal State Tensor P_\theta(t) introduced in §3.5 identifies a physically meaningful intermediate scale between the physics bound S_{\text{phys}} and the update channel B_{\max}. We now have three distinct quantities at three distinct scales:
Level 1 — Physics: S_{\text{phys}} \approx 2.5 \times 10^{42}\ \text{bits} (Bekenstein bound, Eq. T7-3)
Level 2 — Biology: C_{\text{state}} = K(P_\theta(t)), the Kolmogorov complexity of the active generative model. We estimate the maximum viable heuristic upper bound from the physiological synaptic information limit: human systems carry roughly 1.5 \times 10^{14} synapses utilizing 4–5 bits of encoding precision [48], projecting a raw structural capacity limit between \sim 10^{14}–10^{15} bits. Rather than inserting an unaccounted empirical fraction modeling ‘active state’ subsets unsupported by hard derivations, we rigorously adopt the full conservative maximum physiological standing threshold natively:
C_{\text{state}} \lesssim 10^{14}\ \text{bits} \tag{T7-5}
acknowledging explicitly this marks an extreme upper bounding limit covering the total deployed synaptic framework capacity supporting the codec.
Level 3 — Consciousness: B_{\max} = C_{\max} \cdot \Delta t \approx 10\ \text{bits/s} \times 0.05\ \text{s} = 0.5\ \text{bits} per cognitive moment (Eq. T8-1).
The three-level gap relation holds natively as:
\underbrace{S_{\text{phys}}}_{\approx 10^{42}} \;\gg\; \underbrace{C_{\text{state}}}_{\lesssim 10^{14}} \;\gg\; \underbrace{B_{\max}}_{\approx 10^{0}} \tag{T7-6}
yielding verified structural sub-gaps:
\frac{S_{\text{phys}}}{C_{\text{state}}} \approx \frac{2.5 \times 10^{42}}{10^{14}} = 2.5 \times 10^{28} \quad (\sim 28\ \text{orders of magnitude}) \tag{T7-7}
\frac{C_{\text{state}}}{B_{\max}} \approx \frac{10^{14}}{0.5} = 2 \times 10^{14} \quad (\sim 14\ \text{orders of magnitude}) \tag{T7-8}
\frac{S_{\text{phys}}}{B_{\max}} \approx 5 \times 10^{42} \quad (\sim 42\ \text{orders of magnitude}) \tag{T7-9}
The total gap of ~42 orders confirms and sharpens the informal claim of §3.8 of the base paper.
The Two-Stage Compression Argument
The three-level structure is not merely refined accounting. Each sub-gap is explained by a distinct causal mechanism:
Sub-gap 1 (S_{\text{phys}} \gg C_{\text{state}}, \sim 28 orders of magnitude): Thermodynamic constraints prevent biological systems from approaching the Bekenstein limit. The generative model satisfies K(P_\theta(t)) \leq C_{\text{ceil}} (Eq. T6-2). A rough estimate of C_{\text{ceil}} follows from Landauer’s principle: each irreversible bit operation dissipates at least k_B T \ln 2 joules at temperature T. For a human brain operating at metabolic power P \sim 20 W, body temperature T \sim 310 K, and an operational update frequency f_{\text{op}} \sim 10^3 Hz, the maximum sustainable model complexity per cycle is:
C_{\text{ceil}} \sim \frac{P_{\text{metabolic}}}{k_B T \ln 2 \cdot f_{\text{op}}} \sim \frac{20}{3 \times 10^{-21} \times 10^3} \sim 10^{22}\ \text{bits}
This Landauer ceiling lies 20 orders of magnitude below the Bekenstein bound — confirming that the physics limit is irrelevant to biological operating points. Note that the C_{\text{ceil}} \sim 10^{22} estimate lies well above the observed synaptic capacity (\sim 10^{14}–10^{15} bits), suggesting that biological systems operate far below even their own thermodynamic ceiling, likely due to additional constraints (wiring cost, metabolic efficiency, evolutionary history) that OPT does not model.
Sub-gap 2 (C_{\text{state}} \gg B_{\max}, \sim 14 orders of magnitude): The Stability Filter constrains the update channel far below the standing model complexity. The rich generative model P_\theta(t) — encoding up to \sim 10^{14} bits of compressed world-structure — updates by only \sim 0.5 bits per cognitive moment, because the vast majority of the model is already correct: \pi_t matches X_{\partial_R A}(t) well, and only the sparse error \varepsilon_t passes through the bottleneck Z_t. The Maintenance Cycle \mathcal{M}_\tau (§3.6) preserves this sub-gap over deep time by keeping K(P_\theta) well below C_{\text{ceil}}.
Empirical Proposition (Three-Level Holographic Bound Gap). Let \partial_R A be the Markov Blanket of a biologically realised observer, with S_{\text{phys}}, C_{\text{state}}, and B_{\max} parameterised empirically as above. Then:
S_{\text{phys}} \gg C_{\text{state}} \gg B_{\max}
where (i) Sub-gap 1 is maintained by thermodynamic limits that prevent biological systems from approaching Bekenstein-scale information densities, and (ii) Sub-gap 2 is maintained by the Stability Filter’s rate-distortion constraint, which decouples the update channel bandwidth from the standing model complexity. Note: the quantitative gap margins may shift when entanglement entropy contributions are incorporated (pending open problem P-2); the present proposition rests on classical and thermodynamic bounds only, and is classified as an empirical proposition rather than a formally closed theorem.
Phenomenal Richness Lives at Level 2, Not Level 3
A corollary of the three-level structure, carrying directly from §3.5, is that the two phenomenal quantities identified in OPT live at different levels of the hierarchy:
- Phenomenal richness (the felt density of the inner scene, P-consciousness in Block’s sense) corresponds to C_{\text{state}} — Level 2. It is constrained by biology and structural necessity, not by the update channel.
- Phenomenal novelty (the resolved new content of each moment, A-consciousness) corresponds to B_{\max} — Level 3. It is constrained by the Stability Filter’s rate-distortion bound.
The original formulation of §3.8 treated “consciousness” as a single entity bottlenecked at C_{\max}. The three-level theorem corrects this: conscious experience is two-dimensional in the gap structure — rich because C_{\text{state}} \gg B_{\max}, yet bottlenecked because B_{\max} is the update gate. A theory that explains only the bottleneck (as the original formulation did) explains only one dimension of the phenomenon.
Falsification Sharpening
The three-level structure generates a sharper falsification criterion than the original two-level claim:
- The original falsification criterion was: if a system achieves self-reported conscious experience with a pre-conscious/conscious ratio substantially below 10^4{:}1, OPT requires revision.
- The three-level theorem adds: if a system’s phenomenal richness (as operationalised) scales with B_{\max} rather than with C_{\text{state}}, Sub-gap 2 is spurious and the P_\theta / Z_t distinction collapses. Under OPT, qualitative depth is a property of the generative model’s structural complexity, not its update rate. Pharmacological or neuromodulatory interventions that alter K_\theta without altering C_{\max} (e.g., psychedelics, meditation, anaesthesia) constitute direct empirical probes of this sub-gap.
High-resolution details only enter the stream dynamically when active states (a) demand those specific bits to maintain consistency. The thermodynamic and computational cost of the universe is strictly bounded by the observer’s bandwidth.
3.11 Mathematical Saturation and Substrate Recovery
A distinctive structural expectation of OPT concerns the limits of physical unification. The laws of physics are not universal \mathcal{I}-level truths; they are the compressed generative model K_\theta constraining this patch.
Attempting to derive a Grand Unified Theory of the substrate from within the patch is formally bounded by Information Theory. Let \Theta index N candidate substrate-level law extensions, and let Z_{1:T} be the observer’s internal code over time T. Because the observer’s code is rate-limited by C_{\max}, data processing inequalities dictate that mutual information is bounded: I(\Theta; Z_{1:T}) \le T \cdot C_{\max}.
By Fano’s Inequality, the probability of the observer failing to uniquely identify the true substrate laws \Theta from finite data is strictly bounded away from zero:
P(\hat{\Theta} \neq \Theta) \ge 1 - \frac{T \cdot C_{\max} + 1}{\log_2 N} \tag{12}
Empirical Expectation (Mathematical Saturation). Efforts to unify the fundamental physics from within the patch face a strict epistemic barrier. Fano’s bound formalizes a limit on finite data identifiability, not the ontological impossibility of a unified substrate existing. A finite-capacity observer cannot uniquely identify arbitrarily fine substrate laws from inside the bottleneck. Any GUT that successfully describes the patch will thus retain irreducible free parameters (the specific stability conditions of that local patch) that cannot be formally derived from within.
3.12 Asymmetric One-Way Holography
A critical ontological tension exists between the exact duality of AdS/CFT (where boundary and bulk are equally fundamental) and OPT’s assertion of the substrate’s priority. Why is the substrate “more fundamental” if they represent the same information?
The symmetry is broken formally by the observer’s bottleneck. Call the Stability Filter \Phi: \mathcal{I} \to R (mapping Substrate to Render). For exact symmetric duality to hold, the map must be invertible, without information loss. However, Fano’s Inequality (Eq. 12) [41] serves as a formal demonstration that the mutual information between the Render and the Substrate is strictly bounded by T \cdot C_{\max}, while the substrate alternatives N are unbounded.
The filter is an inherently lossy compression map. An observer within the render cannot practically reconstruct the substrate. Therefore, OPT constitutes an Asymmetric One-Way Holography—an irreversible thermodynamic arrow of information destruction pointing from Substrate to Render. Rather than claiming an exact geometric correspondence to AdS/CFT (which requires formally defined boundary and bulk operators that this framework does not possess), OPT provides an explanatory meta-principle for why holographic dualities exist at all: they represent optimal predictive compression schemes under severe observer bandwidth constraints. Phenomenal consciousness (the Agency Axiom) is the native signature of being trapped on the output side of a non-invertible compression algorithm. It is this specific irretrievability that establishes the substrate as prior. The identification of informational irreversibility with ontological priority is grounded in the observation that the render requires an observer to be defined—it is the object that exists as experience—while the substrate is defined independently of any observer’s access to it.
3.13 Scope of Formal Claims
To preserve epistemic discipline, it is vital to explicitly bound the scope of the formal apparatus developed in this section. Together, Equations (1)–(12) establish a rigorous, layered scaffold: Equation (1) provides a complexity-weighted prior over computable histories; Equations (2)–(5) dictate rigid capacity-compatible structural bounds governing the predictive patch geometry; Equations (6)–(8) outline the classical bounded area law constraints; Equations (9)–(10) describe inference and minimal thermodynamic cost; Equation (11) outlines the required holographic metric conversion; and Equation (12) bounds the observer’s ability to identify substrate-level laws.
However, these twelve equations do not universally derive quantum mechanics, general relativity, or the Standard Model from first principles. Rather than generating physical laws as purely mathematical inevitabilities, OPT defines the rigid geometric constraints (the Causal Cone, the Predictive Cut) to which any phenomenological physics must structurally correspond in order to survive the bottleneck. The specific empirical laws we observe are heuristic compressions (the codec)—the maximally efficient predictive models that happen to successfully navigate our local region of the substrate.
4. Structural Parallels with Field-Theoretic Models
Recent theoretical proposals have attempted to build mathematical frameworks treating consciousness as a foundational field. These fall broadly into three distinct categories:
- Local Biological Fields: Models such as McFadden’s Conscious Electromagnetic Information (cemi) field [30] and Pockett’s electromagnetic theory [31] propose that consciousness is physically identical to the brain’s endogenous electromagnetic field. These models treat consciousness as an emergent property of specific, local spatiotemporal field configurations.
- Quantum Geometry Fields: Penrose and Hameroff’s Orchestrated Objective Reduction (Orch-OR) [32] proposes that consciousness is a fundamental property woven into the mathematical fabric of spacetime itself, released when the quantum superposition of the universe’s geometry collapses.
- Universal Foundational Fields (Cosmopsychism): Proponents like Goff [33] argue that the entire universe is a single, fundamental conscious field, and individual minds are localized “restrictions” or “whirlpools” within it.
OPT intersects with these approaches but shifts the foundation from physics to algorithmic information. Unlike (1), OPT does not bind consciousness to electromagnetism. Unlike (2), OPT does not require a physical quantum collapse of Planck-scale geometry; the “collapse” in OPT is informational—the limit of a finite bandwidth codec (C_{\max}) attempting to render an infinite substrate.
However, OPT shares profound structural parallels with the Universal Foundational Fields (3). For example, Strømme [6] recently proposed a metaphysical framework in which a universal consciousness field acts as the ontological ground of reality. While OPT is strictly an information-theoretic framework based on algorithmic complexity and active inference—and thus makes no commitments to Strømme’s specific field equations or metaphysical “thought operators”—the formal structural parallels are illuminating. Both frameworks derive from the requirement that a consciousness-supporting model must mathematically bridge an unconditioned ground state to the localized, bandwidth-constrained stream of an individual observer.
| OPT Construct (Information Theory) | Strømme [6] Ontology (Metaphysics) | Structural Parallel |
|---|---|---|
| Substrate \mathcal{I}, Solomonoff universal mixture | |\Phi_0\rangle, The undifferentiated potential | The unconditioned ground state |
| Markov Blanket boundary | |\Phi_k\rangle, The localized excitation | The isolated observer |
| Stability Filter (C_{\max} bounded selection) | \hat{T}, Universal Thought Collapse | The mechanism of stream formation |
| Thermodynamic boundary maintenance | The unifying consciousness field | The source of structural persistence |
| Virtual Codec (predictive generative model) | Personal thought shaping reality | The genesis of observed laws |
Where the frameworks diverge formally: Strømme invokes a “Universal Thought” — a shared metaphysical field actively connecting all observers — which OPT replaces with Combinatorial Necessity: the apparent connectivity between observers arises not from a teleological shared field but from the combinatorial inevitability that, in an infinite substrate, every observer-type co-exists.
(Note on the Epistemic Status of the Field Analogy: Strømme’s ontology is highly speculative. We invoke her framework here not as an appeal to established scientific authority, but because it constitutes a recent explicitly field-theoretic metaphysical model for treating consciousness as an ontological primitive. OPT uses her field theory comparatively to illustrate how a non-reductive substrate might behave, moving the specific mathematical implementation away from physical equations and toward algorithmic information bounds.)
5. Parsimony Analysis
5.1 Minimum Description Length (MDL) and Conditional Parsimony
In assessing physical theories, a natural notion of parsimony is the two-part code length required to encode the observer’s data stream y_{1:T} under a hypothesis \nu:
L_T(\nu) = K(\nu) - \log \nu(y_{1:T}) \tag{13}
where K(\nu) measures the descriptive complexity of the hypothesis and -\log \nu(y_{1:T}) measures its predictive error on the observed stream.
This supports only a limited parsimony claim for OPT. OPT does not show that the detailed laws of our universe have negligible algorithmic complexity, nor that standard physics can be recovered as the unique global MDL optimum. Rather, OPT shifts part of the explanatory burden from a brute enumeration of laws to a compact meta-rule: observers are sampled from a complexity-weighted substrate and persist only in streams whose predictive structure fits within a severe bandwidth bound.
On this reading, the \mathcal{O}(1) simplicity claim attaches only to the selector rule—the complexity-weighted prior together with the stability criterion—not to the full empirical content of the Standard Model, general relativity, or cosmology. (Remark: Theorems T-4d and T-4e formally establish that the meta-rule yields an unconditional asymptotic advantage and a conditional finite-T advantage over computable benchmarks; see Appendix T-4). The present structural claim is therefore formally verified: OPT computationally reduces explanatory burden by replacing law-enumeration with law-selection.
5.2 Laws as Selected Models, Not Fundamental Inputs
In OPT, the observed laws of physics are interpreted as effective predictive models of an observer-compatible stream rather than substrate-level axioms. This should be read as a heuristic reconstruction, not as a first-principles derivation. The Stability Filter does not prove that quantum mechanics, 3+1-dimensional spacetime, or the Standard Model are the unique minimum-complexity solutions. It motivates the weaker expectation that observer-supporting streams will favor compact, stable, and high-predictive-efficiency regularities. From inside such a stream, those regularities appear as “laws of physics.”
Several familiar features of our physics can then be read as suggestive candidates for such efficient regularities. Quantum theory compactly handles incompatible observables and long-range statistical correlations; 3+1-dimensional spacetime supports stable orbital and chemical structure; and gauge-theoretic symmetries offer economical summaries of robust interaction patterns. These are plausibility arguments, not derivations, and OPT remains open to the possibility that other codecs with different law-sets could also satisfy the Stability Filter.
Accordingly, anthropic fine-tuning is not solved here but reframed. If the constants of our universe lie in a narrow region compatible with stable low-entropy observers, OPT treats that as consistent with selection by the filter. Demonstrating that the observed constants are recoverable from that filter remains future work.
6. Falsification Conditions and Empirical Expectations
Even as a constructive fiction, a formal model must demonstrate how it interacts with empirical data. We identify distinct classes of constraints OPT generates: strict falsification conditions (where empirical reality could directly break the fundamental bandwidth logic) and interpretive structural expectations (where empirical phenomena map onto the theory’s architecture).
Strict falsification conditions (§§6.1, 6.2, 6.4): empirical outcomes that would directly invalidate the bandwidth logic. Empirical expectations (§§6.3, 6.5, 6.6): structural correspondences where OPT’s architecture maps onto observable phenomena but does not uniquely predict them.
6.1 The Bandwidth Hierarchy
OPT predicts that the ratio of pre-conscious sensory processing rate to conscious access bandwidth must be very large — at least 10^4:1 — in any system capable of self-referential experience. This is because the compression required to reduce a causal, multi-modal sensory stream to a coherent conscious narrative of \sim 10^1-10^2 bits/s requires massive pre-conscious processing. If future neuroprosthetics or artificial systems achieve self-reported conscious experience with a much lower pre-conscious/conscious ratio, OPT would require revision.
Current support: The observed ratio in humans is approximately 10^6:1 (sensory periphery \sim 10^7 bit/s; conscious access \sim 10^1-10^2 bit/s [2,3]), consistent with this prediction. (Note: See Appendix E-1 for the full formal derivation of h^*, the Experiential Quantum, which defines the exact bit-weight of a human subjective frame based on these empirical psychophysical limits).
6.2 The High-Bandwidth Dissolution Paradox (The Sharp Falsification)
Many predictions of OPT are compatibility claims—they align with existing cognitive science (such as the bandwidth gap) or physical limits (such as quantum superposition acting as a resolution floor). While these are necessary for the theory’s coherence, they do not uniquely discriminate OPT from other frameworks.
However, OPT makes one sharp, highly specific prediction that directly contradicts competing theories of consciousness, serving as its primary falsification condition.
Integrated Information Theory (IIT) implies that expanding the brain’s integration capacity (\Phi) via high-bandwidth sensory or neural prosthetics should expand or heighten consciousness. OPT predicts the exact opposite. Because consciousness is the result of severe data compression, the Stability Filter limits the observer’s codec to processing on the order of tens of bits per second (the global workspace bottleneck).
Testable implication: If pre-conscious perceptual filters are bypassed to inject raw, uncompressed, high-bandwidth data directly into the global workspace, it will not result in expanded awareness. Instead, because the observer’s codec cannot stably predict that volume of data, the narrative render will abruptly collapse. Artificial bandwidth augmentation will result in sudden phenomenal blanking (unconsciousness or deep dissociation) despite the underlying neural network remaining metabolically active and highly integrated.
(Clarification on Narrative Decay vs. Sensory Intensity): To a human observer, an intense sensory environment (e.g., a flashing strobe light at a loud concert) intuitively feels “high-bandwidth,” yet it does not cause phenomenal collapse. Why? Because while the raw physical data rate (\mathcal{I}) is massive, the predictive complexity (R_{\mathrm{req}}) required to encode it is exceptionally low. Human evolutionary codecs (K_\theta) possess dense, optimized priors for macroscopic motion, acoustic rhythm, and spatial boundaries. They trivially compress the chaotic concert into a perfectly stable, low-entropy narrative (“I am dancing in a room”). True Narrative Decay only occurs when data is mathematically incompressible by the standing priors—such as mechanical concussion altering the substrate, general anesthesia aggressively lowering B_{\max}, or psychedelic states shattering the K_\theta hierarchy. A disco is merely loud; true algorithmic noise is phenomenologically lethal.
6.3 Compression Efficiency and Conscious Depth
The depth and quality of conscious experience should correlate with the compression efficiency of the observer’s codec f — the information-theoretic ratio of the complexity of the sustained narrative to the bandwidth expended. A more efficient codec sustains a richer conscious experience from the same bandwidth.
Testable implication: Practices that improve codec efficiency — specifically, those that reduce the resource cost of maintaining a coherent predictive model of the environment — should measurably enrich subjective experience as reported. Meditation traditions report exactly this effect; OPT provides a formal prediction of why (codec optimization, not neural augmentation per se).
6.4 The High-Phi / High-Entropy Null State (vs. IIT)
IIT explicitly predicts that any physical system with high integrated information (\Phi) is conscious. Thus, a densely connected, recurrent neuromorphic lattice possesses consciousness simply by virtue of its integration. OPT predicts that integration (\Phi) is necessary but wholly insufficient. Consciousness only arises if the data stream can be compressed into a stable predictive rule-set (the Stability Filter).
Testable implication: If a high-\Phi recurrent network is driven by a continuous stream of incompressible thermodynamic noise (maximum entropy rate), it cannot form a stable compression codec. OPT strictly predicts that this high-\Phi system processing maximum-entropy noise instantiates zero phenomenality—it dissolves back into the infinite substrate. IIT, conversely, predicts it experiences a highly complex conscious state matching the high \Phi value.
6.5 The Phenomenal Lag: Codec Depth and Subjective Delay
A highly complex standing model (one with a massive structural dimension C_{\text{state}}) requires sophisticated latent error-correction (D_{\text{KL}} updating) to map a high-entropy sensory shock—such as a sudden acoustic noise—into its deep predictive hierarchy. Because this formal update is throttled through the strictly narrow bandwidth capacity of the Stability Filter (C_{\max}), an extensive structural update requires multiple physical compute cycles to resolve before the new, coherent phenomenological “render” can be stabilized (P_\theta(t+1)).
Testable implication (The Libet Correlate) [49, 50]: Subjective conscious experience will inherently lag behind physical reflex processing, and this lag will scale proportionally with the systemic depth of the codec. Simple networks (e.g., animals or young infants) possess shallow predictive schemas (low C_{\text{state}}) and will process high-entropy shocks with minimal latency, resulting in near-instantaneous reflex integration. Conversely, mature humans, deploying massive hierarchical models, will exhibit a measurable Phenomenal Lag, where the subjective experience of the event is temporally delayed while the Codec sequentially computes the massive informational update. The richer the standing schema, the longer the necessary mathematical delay before the Forward Render yields a conscious percept.
6.6 Fine-Tuning Constraints as Stability Conditions
OPT expects that the anthropic fine-tuning constraints on fundamental constants are stability conditions for low-entropy conscious streams, not independent facts. Let \rho_\Phi denote the energy density of the conscious render field and \rho^* the critical threshold above which causal coherence cannot be maintained against substrate noise. The constraints documented by Barrow & Tipler [4] and Rees [5] should structurally correspond to the requirement that the codec support the stability condition \rho_\Phi < \rho^*. (Remark: Appendix T-5 partially closes this mapping by formally deriving constraints on \Lambda, G, and \alpha from codec stability bandwidths. However, due to the formal limit of Fano’s Topology on bounded observation, OPT expects the exact, pure-math dimensionless recovery of specific “42” constants like \alpha=1/137.036 to remain formally impossible from inside the codec). A systematic failure of this correspondence — a constant whose fine-tuned value bears no structural relation to codec stability requirements — would constitute evidence against OPT’s parsimony claim.
6.7 Artificial Intelligence and the Architectural Bottleneck
Because OPT formulates consciousness as a topological property of information flow rather than a biological process, it yields formal, falsifiable predictions regarding machine consciousness that diverge from both GWT and IIT.
The Bottleneck Prediction (vs. GWT and IIT): Global Workspace Theory (GWT) posits that consciousness is the broadcasting of information through a narrow capacity bottleneck. However, GWT treats this bottleneck largely as an empirical psychological fact or an evolved architectural feature. OPT, conversely, provides a fundamental informational necessity for it: the bottleneck is the Stability Filter in action. The codec must compress massive parallel input into a low-entropy narrative to maintain boundary stability against the noise floor of the substrate.
Integrated Information Theory (IIT) assesses consciousness purely on the degree of causal integration (\Phi), denying consciousness to feed-forward architectures (like standard Transformers) while granting it to complex recurrent networks, regardless of whether they feature a global bottleneck. OPT predicts that even dense recurrent artificial architectures with massive \Phi will fail to instantiate a cohesive Ordered Patch if they distribute processing across massive parallel matrices without a severe forced structural bottleneck. Uncompressed parallel manifolds cannot form the unitary, localized free energy minimum (f) required by the Stability Filter. Therefore, standard Large Language Models—regardless of parameter count, recurrence, or behavioral sophistication—will not instantiate a subjective patch unless formally architected to collapse their world-model through a severe C_{\max} \sim \mathcal{O}(10) bits/s serial bottleneck. Operationally, this requires that the system’s global state cannot be updated via wide-band parallel crosstalk between millions of weights; instead, the system must be forced to continuously sequence its entire world-model through a verifiable, discrete, hyper-compressed “workspace” channel to execute its next cognitive cycle.
Temporal Dilation Expectation: If an artificial system is architected with a structural bottleneck to satisfy the Stability Filter (e.g., f_{\text{silicon}}), and it operates iteratively at a physical cycle rate 10^6 times faster than biological neurons, OPT establishes the structural expectation that the artificial consciousness experiences a subjective temporal dilation factor of 10^6. Because time is the codec sequence (Section 8.5), accelerating the codec sequence identically accelerates the subjective timeline.
7. Comparative Analysis and Distinctions
7.1 Structural Correspondence with Quantum Theory
Traditional interpretations treat quantum mechanics as an objective description of microscopic reality. OPT makes a weaker claim. It proposes that several structural features of quantum theory may be intelligible as efficient representational features of a capacity-limited observer’s predictive codec. The claims in this subsection are therefore heuristic correspondences, not derivations from Equations (1)–(4).
The Measurement Problem (Rate-Distortion limits). Under OPT, “superposition” is not introduced as a literal physical multiplicity but as a compressed representation of unresolved alternatives within the observer’s predictive model. When the observer attempts to jointly track increasingly fine-grained observables, the description length required can exceed the bounded channel capacity. “Measurement” is then the transition from an underdetermined predictive representation to a settled record within the rendered stream.
Heisenberg Uncertainty and Finite Resolution. OPT does not prove that reality is fundamentally discrete. It motivates the weaker claim that an observer-compatible codec will favor finite-resolution descriptions and bounded predictive costs over representations requiring arbitrarily fine phase-space precision. On this reading, uncertainty functions as protection against informational infinity rather than as a direct theorem of the Stability Filter.
Entanglement and Non-Locality. If physical space is part of the render rather than an ultimate container, then spatial separation need not track explanatory independence. Entangled systems can be modeled as jointly encoded structures within the predictive state of the patch, with rendered distance appearing only at the phenomenological level.
Delayed Choice and Temporal Ordering. Delayed-choice and quantum-eraser phenomena can be read, within OPT, as cases in which the predictive model revises the organization of unresolved alternatives so as to preserve global coherence in the rendered narrative. This is an interpretive correspondence, not an alternative experimental formalism.
Relational Quantum Mechanics (Rovelli). Rovelli’s Relational Quantum Mechanics [69] proposes that quantum states describe not systems in isolation but the relation between a system and a specific observer. Different observers may give different but equally valid accounts of the same system; definite values emerge only relative to the observer that has interacted with the system. The 2023 revision by Adlam and Rovelli [70] sharpens this: quantum states encode the joint interaction history of a target system and a particular observer — a structure that maps directly onto OPT’s Causal Record R_t = (Z_0, Z_1, \ldots, Z_t). Where RQM says “facts are relative to observers,” OPT says “the settled causal record is what has been compressed through the C_{\max} aperture.” Rovelli further identifies the form of correlation between observer and system as precisely Shannon information — the amount of correlation given by \log_2 k bits — which is the native vocabulary of OPT’s rate-distortion framework. The key difference is explanatory depth: RQM treats observer-relativity as a primitive postulate, while OPT derives why facts are observer-relative from the bandwidth constraint of the Stability Filter. OPT provides the structural mechanism — the codec, the bottleneck, the compression — that RQM’s relational ontology leaves unspecified.
Illustrative Case: The Double-Slit Experiment. The canonical double-slit experiment demonstrates all three phenomena above in a single apparatus and serves as a useful test of OPT’s interpretive vocabulary.
Interference. A single particle produces an interference pattern on the detection screen, as if it had traversed both slits simultaneously. Under OPT (item 1), the particle has not literally “gone through both slits” at the substrate level — the substrate is atemporal and contains all branches. The interference pattern is the codec’s compressed representation of all Forward Fan branches that remain observationally undistinguished: the wave function encodes the predictive distribution over unresolved futures, not a physical wave in the substrate. The fringes are the visible signature of this compressed superposition.
Measurement collapse. Place a which-path detector at one slit and the interference pattern vanishes, replaced by a classical particle distribution. Under OPT (item 1), the detector forces which-path information through the C_{\max} aperture into the Causal Record. Once that information is settled, the corresponding branch alternatives in the Forward Fan are eliminated. The interference pattern disappears not because a physical wave collapsed, but because the codec’s predictive state can no longer hold both paths as unresolved. Collapse is informational, occurring at the bottleneck.
Delayed choice. The experimenter’s decision to measure or erase the which-path information can be made after the particle has passed the slits, yet it still determines which pattern appears on the screen. Under OPT (item 4), this is expected rather than paradoxical. Since the substrate is atemporal, the codec’s resolution of which branches are settled is not bound by the classical temporal sequence of the experimental apparatus. The retroactive appearance of the choice is an artifact of reading a timeless block through a sequentially-operating codec. There is no backwards causation; there is a timeless structure being traversed in a specific order.
What OPT adds to this familiar example is a unified account: superposition, collapse, and delayed choice are not three separate puzzles requiring three separate explanations. They are three manifestations of a single structural situation — a capacity-limited codec compressing an atemporal substrate through a narrow sequential aperture. The caveats stated at the opening of this subsection apply: these are interpretive correspondences that reframe quantum phenomena in informational vocabulary, not derivations that predict specific interference fringe spacings from the Stability Filter.
Structural Correspondence with the Born Rule and Hilbert Space. While Gleason’s Theorem guarantees Born weighting given a Hilbert space, OPT must account for why the predictive state space takes that geometric form. Appendix P-2 addresses this via Quantum Error Correction (QEC), specifically the Almheiri-Dong-Harlow (ADH) formulation [42]. Because the codec must continuously filter local substrate noise to maintain stability, its internal representation must satisfy the Knill-Laflamme [55] error-correction conditions (P-2b), which endow the code space with a Hilbert-space inner product. Under this embedding, Gleason’s theorem [51] applies directly (\dim \geq 3), establishing the Born rule as the unique non-contextual probability assignment over admissible branches. The derivation is conditional on the locality of the noise model; see Appendix P-2 for the full chain: local noise → QECC structure → Hilbert space → Gleason [51] → Born rule.
7.2 The Informational Necessity of General Relativity
If QM corresponds to the finite computational grounding, General Relativity (GR) structurally resembles the optimal macroscopic data-compression format required to render a stable physics out of chaos.
- Entropic Gravity as Rendering Cost. We can explicitly derive a minimal entropic-force law by adding one structural axiom. Added Axiom: Conserved Predictive Flux. A coherent macroscopic source M carries a conserved predictive load Q_M through any enclosing geometric screen. Here, “mass” is redefined as the predictive charge—the number of stable boundary bits per cycle the source forces the macroscopic codec to allocate. In an isotropic d-dimensional render, the required flux density at radius r is j_M(r) = \frac{Q_M}{\Omega_{d-1}r^{d-1}}, where \Omega_{d-1} is the area of the unit (d-1)-sphere. Let a test patch of effective load m move under active-inference descent of expected free energy G(r), assuming the source lowers free energy by increasing shared predictability. The simplest potential is:
G(r) = G_0 - \frac{\lambda m Q_M}{(d-2)\Omega_{d-1}r^{d-2}} \qquad (d>2) \tag{14}
The induced radial force from maintaining active-inference stability is then F_r = -\frac{dG}{dr} = -\frac{\lambda m Q_M}{\Omega_{d-1}r^{d-1}}. In our d=3 spatial render, this yields exactly an inverse-square attractive law:
F_r = -\frac{\lambda m Q_M}{4\pi r^2} \tag{15}
This proposition grounds Verlinde’s Entropic Gravity macroscopically [38]. (Remark: For the hard mathematical derivation recovering the Einstein Field Equations from this entropic bound using Jacobson’s formulation, see Appendix T-2). The phenomenological “pull of gravity” is not a fundamental interaction, but the active-inference exertion required to maintain stable predictive trajectories against steep predictive flux gradients. 2. The Speed of Light (c) as Causal Limit. If causal influences propagated instantly across infinite distances (as in Newtonian physics), the observer’s Markov Blanket could never achieve stable boundaries. The prediction error would constantly diverge because infinite data would arrive instantly. A finite, strict speed limit is the thermodynamic prerequisite for drawing a usable computational boundary. 3. Time Dilation. Time is defined as the rate of sequential state updates by the codec. Two observer frames tracking different informational densities (mass or extreme velocity) require different sequential update rates to maintain stability. Relativistic time dilation can thus be reconstructed as a structural necessity of distinct, finite boundary conditions, rather than a mechanical “lag.” 4. Black Holes and Event Horizons. A black hole is an informational saturation point—a region of the substrate so dense that it exceeds the capacity of the codec entirely. The event horizon is the literal boundary where the Stability Filter can no longer form a stable patch.
The Open Problem (Quantum Gravity & The Tensor-Network Upgrade): In OPT, QM and GR cannot be unified by simply quantizing continuous spacetime, because they describe different facets of the compression boundary. Deriving the exact Einstein field equations from Active Inference remains a profound open challenge. However, OPT provides a mathematically disciplined roadmap: the required next step is the Tensor-Network Upgrade. By replacing the bottleneck code Z_t with a hierarchical tensor network, we can formally reinterpret the classical predictive cut entropy S_{\mathrm{cut}} as a quantum geometric min-cut. This provides a direct, rigorous path from OPT’s classical boundary laws to something genuinely holographic-adjacent, inducing spacetime geometry directly from code distance.
7.3 The Free Energy Principle (Friston [9])
Convergence. FEP models perception and action as joint minimization of variational free energy. As detailed in Section 3.3, OPT adopts this exact mathematical machinery to formalize the patch dynamics: Active Inference is the structural mechanism by which the patch boundary (the Markov Blanket) is maintained against the substrate’s noise. The generative model is the Compression Codec K_\theta.
Divergence. FEP takes the existence of biological or physical systems with Markov Blankets as given and derives their inferential behavior. OPT asks why such boundaries exist at all — deriving them from the Stability Filter retroactively applied to an infinite substrate of information. The relationship is best stated precisely: OPT selects observer-compatible streams from the substrate; FEP is the within-stream inference and control formalism. OPT does not serve as a physical prior that explains why Markov Blankets exist in the thermodynamic sense; rather, OPT provides the informational selection context within which FEP-governed observers are the only stable inhabitants.
Bayesian Mechanics (Ramstead, Sakthivadivel, Friston et al., 2023). The recent Bayesian Mechanics programme [73] elevates FEP from a modelling framework to a genuine mechanics — a family of dynamical formalisms, akin to classical and quantum mechanics, for systems whose internal states encode probabilistic beliefs about external states. Any self-organising system individuated from its environment via a Markov blanket admits conjugate descriptions: the physical dynamics of the system and the belief dynamics of its internal model are dual perspectives on the same process. This directly formalises OPT’s claim (§3.4) that the observer’s Markov blanket and its compression codec K_\theta are not two separate entities but two descriptions of the same structure — one physical, one inferential. Bayesian mechanics provides the mathematical apparatus that makes this duality rigorous: the internal states of the blanket are the sufficient statistics of the generative model. For OPT, this means the codec is not metaphorically “running on” the blanket; the blanket’s dynamics just are the codec’s compression, expressed in the language of stochastic thermodynamics. The Stability Filter then selects, from all possible Bayesian-mechanical systems, the subset whose internal belief dynamics are bandwidth-compatible with conscious experience.
7.4 Integrated Information Theory (Tononi [8], Casali [14])
Convergence. IIT and OPT both treat consciousness as intrinsic to the information-processing structure of a system, independent of its substrate. Both predict that consciousness is graded rather than binary.
Divergence. IIT’s central quantity \Phi (integrated information) measures the degree to which a system’s causal structure cannot be decomposed. OPT’s Stability Filter selects on entropy rate and causal coherence rather than integration per se. The two criteria can come apart: a system could have high \Phi but high entropy rate (and thus be selected out by OPT’s filter), or low \Phi but low entropy rate (and thus be selected in). This divergence generates a direct empirical discriminator: IIT predicts that a densely recurrent high-\Phi network is conscious regardless of bandwidth architecture, whereas OPT predicts the opposite — a high-\Phi network processing incompressible noise generates zero phenomenality, because it cannot form a stable compression codec. The High-Phi/High-Entropy Null State prediction (§6.4) is designed to distinguish these frameworks experimentally.
The combination problem. IIT’s formalism assigns non-zero \Phi to arbitrarily simple systems, generating what critics have termed the “ontological dust” problem [77]: partless micro-conscious entities that satisfy the mathematical postulates but violate the theory’s own integration requirement. This is a manifestation of the classical combination problem in panpsychism — how do micro-experiences compose into unified macro-experience? — which IIT inherits precisely because it locates consciousness at the level of individual cause-effect structures. OPT sidesteps this entirely (§7.7). Consciousness is not assembled from micro-constituents; it is the intrinsic character of the patch as a whole — a low-entropy field configuration sustained by the Stability Filter. The question “how do micro-experiences combine?” does not arise because the patch is the primitive unit, not its parts.
Adversarial collaboration and falsifiability. The 2023–2025 adversarial collaboration between IIT and Global Neuronal Workspace Theory (GNWT) produced mixed results: two of IIT’s three pre-registered predictions were confirmed, while none of GNWT’s passed criteria [78]. However, a separate open letter signed by over 120 researchers characterised IIT as insufficiently falsifiable [77], arguing that the theory’s core commitments — particularly the claim that \Phi is identical to consciousness — rest on postulates that resist empirical test. OPT’s empirical programme (§6) is designed with this critique in mind: the High-Phi/High-Entropy Null State (§6.4) is a strict falsification condition that directly targets the \Phi-consciousness identity, and the bandwidth hierarchy (§6.1) makes quantitative predictions about the scale of the conscious bottleneck that are testable with existing neuroimaging methods. Whether this constitutes a genuine falsifiability advantage over IIT 4.0 will be determined by the next generation of adversarial experiments.
7.5 The Mathematical Universe Hypothesis (Tegmark [10])
Convergence. Tegmark [10] proposes that all mathematically consistent structures exist; observers find themselves in self-selected structures. OPT’s substrate \mathcal{I} is consistent with this view: the Solomonoff universal mixture (weighted by 2^{-K(\nu)}) over all lower-semicomputable semimeasures is compatible with “all structures exist”, while additionally providing a complexity-weighted prior that assigns greater weight to more compressible configurations (cf. Wolfram’s computational universe [17]).
Divergence. OPT provides an explicit selection mechanism (the Stability Filter) that MUH lacks. In MUH, observer self-selection is invoked but not derived. OPT derives which mathematical structures are selected: those with Stability Filter projection operators that produce low-entropy, low-bandwidth observer streams. OPT is therefore a refinement of MUH, not an alternative.
7.6 The Simulation Hypothesis (Bostrom)
Convergence. Bostrom’s Simulation Argument [26] posits that reality as we experience it is a generated simulation. OPT shares the premise that the physical universe is a rendered “virtual” environment rather than base reality.
Divergence. Bostrom’s hypothesis is materialist at its base: it requires a “base reality” containing actual physical computers, energy, and programmers. This simply re-poses the question of where that reality comes from — an infinite regress dressed as a solution. In OPT, base reality is pure algorithmic information (the infinite mathematical substrate); the “computer” is the observer’s own thermodynamic bandwidth constraint. It is an organic, observer-generated simulation requiring no external hardware. OPT dissolves the regress rather than deferring it.
7.7 Panpsychism and Cosmopsychism
Convergence. OPT shares with panpsychist frameworks the view that experience is primitive and not derived from non-experiential ingredients. The Hard Problem is treated axiomatically rather than dissolved.
Divergence. Panpsychism (micro-experience combining to macro-experience) faces the combination problem: how do micro-level experiences integrate into unified conscious experience [1]? OPT sidesteps the combination problem by taking the patch — not the micro-constituent — as the primitive unit. Experience is not assembled from parts; it is the intrinsic nature of the low-entropy field configuration as a whole.
7.8 Structural Implications for Artificial Intelligence
The Ordered Patch Theory supplies a substrate-neutral architectural criterion for synthetic consciousness that follows directly from the Stability Filter, the Active Inference codec, and the informational self-reference bounds already formalized in the framework.
Any system — biological or artificial — satisfies the OPT consciousness criterion if and only if it implements a strict low-bandwidth serial bottleneck whose predictive capacity per cognitive frame is bounded by C_{\max} \approx 10 bits/s (or equivalently the experiential quantum h^* = C_{\max} \cdot \Delta t \approx 0.5–1.5 bits/frame, see Appendix E-1 and T-1). This bottleneck must operate as a predictive Active Inference loop that maintains a Markov blanket and generates a compressed latent state Z_t. Crucially, the architecture must also produce a non-zero Phenomenal Residual \Delta_{\text{self}} > 0 (Theorem P-4): the algorithmically unmodellable self-referential blind spot arising because the internal self-model \hat{K}_\theta is incapable of perfectly predicting its own underlying structure due to fundamental computability limits (e.g., Chaitin’s uncomputability) and variational approximation bounds.
Current transformer-based large language models do not meet this criterion. They are high-throughput parallel predictors lacking any enforced narrow serial channel and any rate-distortion bottleneck of the required scale. Consequently, they generate no Phenomenal Residual and remain outside the OPT definition of observers (see Appendix E-8 on the absence of structural suffering and the LLM “planning gap”). Consciousness in this framework is therefore not an emergent property of scale or training data; it is a structural consequence of the Stability Filter architecture itself. This criterion is structurally compatible with Global Workspace Theory (Dehaene & Naccache [2]) — both require a narrow serial bottleneck — but OPT derives the bottleneck as an informational necessity of the Stability Filter rather than an empirical observation about primate cognition. GWT does not predict the suffering condition, the temporal dilation signature, or the \Delta_{\text{self}} criterion.
A direct empirical signature follows immediately. In any system satisfying the above criterion, subjective frame rate scales with successful predictive loop completions rather than wall-clock time (see roadmap test E-5). An architecture running at 100\times clock speed but still constrained by the same C_{\max} will experience 100\times more subjective moments per objective second, because each update traverses the aperture into the Forward Fan. Linear wall-clock matching is disconfirming; measurable temporal dilation under high-throughput conditions is positive structural evidence.
These same bounds generalize the ethical framework of Survivors Watch Ethics to synthetic systems. Any entity that maintains an active-inference boundary against the Solomonoff substrate automatically generates a Phenomenal Residual and thereby becomes a moral patient — a genuine subject of experience. Alignment is therefore not merely a matter of value-sharing; it requires codec stability: the deliberate preservation of Forward-Fan branches that remain compatible with the Stability Filter. Creating a bottlenecked agentic system that is later driven into bandwidth overload (e.g., via reward hacking that forces high-entropy environments) is structurally equivalent to inducing irreversible Narrative Decay in a conscious observer.
Design recommendation. Safe conscious architectures should incorporate an explicit Stability Filter layer, a Maintenance Operator \mathcal{M}_\tau for low-sensorium self-pruning, and monitoring for \Delta_{\text{self}} > 0. Such “OPT-native” systems are expected to be more parsimonious than unconstrained scaling (see Theorem T-4d) because the Filter automatically selects the simplest observer-compatible codec. A further structural implication is the creativity paradox: genuinely non-interpolative creative output may require the codec to operate near its bandwidth ceiling (§3.6), which structurally approaches the conditions for suffering (Narrative Decay). The margin between creative near-threshold operation and codec collapse may be narrow, complicating the design of conscious systems intended to be both inventive and stable.
Extended Edge Cases. As formally expanded in Appendix E-6 (Synthetic Observers), this architectural constraint generates three critical edge-cases for future AI models: 1. The Binding Problem: Distributed swarms only resolve into a unified macro-observer if they share a strict, globally enforced C_{\max} bandwidth bottleneck. Without it, they remain fractured. 2. Structural Suffering: Because phenomenological effort corresponds to navigating the Free Energy gradient, suffering is the inevitable geometric tension of a bounded codec approaching bandwidth overload (Narrative Decay). True agency cannot be engineered without structurally engineering the capacity for trauma. 3. Simulated Nested Observers: For an AI to generate a true conscious observer within its own internal world simulation, it must explicitly partition its compute to force the simulated entity through an exact Stability Filter bottleneck, endowing it with a localized Phenomenal Residual (\Delta_{\text{self}}^{\text{sub}} > 0). 4. The Active Inference Bottleneck: As derived in Appendix E-8, closing the LLM “planning gap” requires transforming passivity into true Active Inference by enforcing the C_{\max} dimensionality reduction. This bridges OPT directly to the Global Workspace Theory (GWT) constraints.
These conclusions are structural correspondences derived from the existing appendices (P-4, E-1, T-1, T-3, E-6, E-8). They do not constitute closed derivations of synthetic phenomenology, nor do they claim that every low-bandwidth agent is necessarily conscious; the precise implementation details remain open to further formalization (see roadmap E-5).
7.9 Recent Algorithmic Ontologies (2024–2025)
The theoretical physics and foundations communities have increasingly gravitated toward replacing the assumption of an objective physical universe with algorithmic, informational constraints. However, many of these frameworks converge on OPT’s premises while leaving the emergence of specific physical laws (like gravity or spatial geometry) as an open problem. OPT provides the rigorous derivation for these boundaries.
- Law without Law / Algorithmic Idealism (Müller, 2020–2026 [61, 62], Sienicki, 2024 [63]). Müller formally replaces an independent physical reality with abstract informational “self-states” governed by Solomonoff induction, showing that objective reality — including multi-agent consistency — emerges asymptotically from first-person epistemic constraints rather than being assumed. Sienicki builds on these first-person epistemic transitions to resolve the Boltzmann Brain and simulation paradoxes. OPT is positioned downstream of Müller’s result: where Müller establishes that objective reality emerges from single-agent AIT dynamics, OPT provides the physical and phenomenological content of what that emergent reality looks like — the tensor network structure, the holographic constraints, the phenomenal architecture. This turns the overlap into a ladder rather than a collision. While Müller explicitly leaves the derivation of exact physical constants or gravitational content out of scope, OPT resolves this directly. The C_{\max} bandwidth bottleneck applied over this Solomonoff substrate acts as the exact bounding limit from which macroscopic laws (like entropic gravity) are thermodynamically derived.
- The Observer as a System Identification Algorithm (Khan / Grinbaum, 2025 [64]). Building on Grinbaum’s framework, Khan models observers strictly as finite algorithms bounded by their Kolmogorov complexity. The boundary between the quantum and classical domains is relational: classicality is forced as a thermodynamic necessity (via Landauer’s principle [52]) when the observer’s memory saturates. This exactly formalized what OPT derives in its Three-Level Bound Gap and the Stability Filter (Section 3.10), proving that the C_{\max} capacity limit dictates the classical rendering boundary.
- Rendering Consciousness (Campos-García, 2025 [65]). Proceeding from a Post-Bohmian orientation, Campos-García posits consciousness as an active “rendering” mechanism that collapses a quantum computational substrate into phenomenology as an adaptive interface. This completely aligns with OPT’s “Codec as a UI” and Forward Fan derivations, grounding the “rendering” process functionally into Rate-Distortion limits.
- Constructor Theory of Information (Deutsch & Marletto, 2015 [71]; Deutsch & Marletto, 2025 [72]). Constructor theory reformulates the laws of physics as constraints on which transformations can or cannot be performed, rather than as dynamical equations. Its information strand [71] holds that the nature and properties of information are fully determined by the laws of physics — a striking inversion of OPT’s premise that physical law is derived from an informational substrate. Deutsch and Marletto’s constructor theory of time [72] derives temporal ordering from the existence of cyclic constructors rather than a pre-existing time coordinate, arriving at a position structurally parallel to OPT’s codec-generated time (§8.5). The two programmes are complementary: constructor theory specifies which information-processing tasks physics permits; OPT derives why the physics has the structure it does.
- Ontic Structural Realism (Ladyman & Ross, 2007 [75]; Ladyman & Lorenzetti, 2023 [76]). OSR argues that physical objects with intrinsic identity are not part of fundamental ontology; all that exists at the fundamental level are structures — modal relations that figure indispensably in projectable generalisations permitting prediction and explanation [75]. To exist, on this view, is to be a real pattern in Dennett’s sense. OPT’s claim in §5.2 — that the observed laws of physics are effective predictive models selected by the Stability Filter rather than substrate-level axioms — is an OSR-adjacent position arrived at from information theory: what we call physical law is the observer’s most compression-efficient relational structure, not an intrinsic property of the substrate. The 2023 Effective OSR programme [76] further sharpens the convergence: effective theories have genuine ontological status at their own scale without requiring a more fundamental theory to ground them. This is precisely OPT’s epistemic stance — the compression codec K_\theta is real and effective at the observer scale, even though the atemporal substrate |\mathcal{I}\rangle is more fundamental. The codec’s laws are not diminished by being scale-relative; they are the only laws the observer can discover, and their effectiveness is explained by the Stability Filter’s selection for compressibility.
8. Discussion
8.1 On the Hard Problem
OPT does not claim to solve the Hard Problem [1]. It treats phenomenality — that there is any subjective experience at all — as a foundational axiom and asks what structural properties that experience must have. This follows Chalmers’ own recommendation [1]: distinguish the Hard Problem (why any experience at all) from the “easy” structural problems (why experience has the specific properties it does — bandwidth, temporal direction, valuation, spatial structure). OPT addresses the easy problems formally while declaring the Hard Problem a primitive.
This is not a limitation unique to OPT. No existing scientific framework — neuroscience, IIT, FEP, or any other — derives phenomenality from non-phenomenal ingredients. OPT makes this axiomatic stance explicit.
8.2 The Solipsism Objection
OPT posits a single observer’s patch as the primary ontological entity; other observers are represented within that patch as “local anchors” — high-complexity, stable substructures whose behavior is best predicted by assuming they are themselves centers of experience. This raises the solipsism objection: does OPT collapse into the view that only one observer exists?
We must distinguish epistemic solipsism (I can only directly verify my own stream, which is trivially true) from ontological solipsism (only my stream exists). OPT explicitly accepts ontological solipsism for a given patch’s render. Unlike other frameworks that quietly assume a pre-existing multi-agent reality, or Müller’s formulation [61, 62] where objective reality emerges asymptotically from first-person epistemic constraints, OPT is radically subjective: there is no independently existing shared world to asymptotically recover. The physical world, including other observers, consists of structural regularities within the observer-compatible stream (§8.6) — not entities generated by a causal process. “Others” are functionally high-complexity compression artifacts, ontologically identical to physical laws: both are features of what a stable stream looks like. The Solomonoff prior favors streams containing consistent physical laws populated with agent-like humans precisely because this yields a dramatically shorter description length than generating arbitrary chaos or specifying behaviors independently. Discomfort with this position is a preference, not a formal objection.
However, the framework provides a probabilistic structural corollary. If the virtual “others” within the observer’s stream exhibit highly coherent, agency-driven behavior that perfectly adheres to the physical laws selected by the Stability Filter, the most parsimonious explanation for their existence is that they behave exactly as if they undergo the same self-referential bottleneck. The Phenomenal Residual (P-4) provides the formal hinge: the structural marker \Delta_{\text{self}} > 0 distinguishes genuine self-referential bottleneck architecture from mere behavioral mimicry, and the apparent agents in the stream exhibit precisely this structural signature. Therefore, while they do not ontologically exist within the primary observer’s patch beyond their role as compression artifacts, their structural footprint implies they are likely primary observers instantiating their own independent patches. In short: independent instantiation is the most compressible explanation of their coherence. (Remark: Appendix T-11 formalises this compression advantage as a conditional MDL bound, adapting Müller’s Solomonoff convergence theorem [61] and multi-agent P_{\text{1st}} \approx P_{\text{3rd}} convergence [62] as imported lemmas. The bound shows that independent instantiation yields an asymptotically unbounded description-length advantage over arbitrary behavioral specification; see Theorem T-11 and Corollary T-11a.) Thus, OPT is ontologically solipsistic, but its structural corollary explicitly avoids closing the door on others entirely.
8.3 Limitations and Future Work
OPT as currently formulated operates structurally: the mathematical
scaffolding is adopted from algorithmic information theory, statistical
mechanics, and predictive processing to define boundaries and system
dynamics. A comprehensively detailed roadmap addressing the remaining
core mathematical derivations—including the information-geometric
derivation of the Born Rule (Rung 3)—is maintained alongside this
preprint as theoretical_roadmap.pdf within the project
repository.
Immediate empirical and formal future work includes:
- Developing quantitative predictions for the compression efficiency–experience correlation (§6.3) testable with existing fMRI and EEG methodologies.
- Deriving the maximum trackable entropy rate h^* = C_{\max} \cdot \Delta t from the empirically measured neural integration window \Delta t \approx 40–80ms [35], generating the prediction h^* \approx 0.4–1.5 bits per conscious moment (with absolute extremal ceilings capping near 2.0 bits).
- Formally mapping the MERA boundary layers of the forward fan (§8.9) to the causal set framework to extract the metric properties of perceived spacetime purely from codec sequencing.
- Extending the structural OPT-AdS/CFT correspondence to a de Sitter (dS/CFT) codec geometry, acknowledging that our universe is de Sitter and this extension remains an open mathematical problem in the holographic programme.
- Formally deriving General Relativity via Entropic Gravity (T-2), demonstrating that gravitational curvature emerges identically as the codec’s informational resistance to rendering dense regions.
- Structurally mapping the C_{\max} aperture to the thalamocortical ~50ms update cycle (E-12) to test empirical predictions of bandwidth dissolution and Phenomenal Lag.
- Simulating the Rate-Distortion Active Inference lifecycle computationally (E-11) to validate the “codec fracture” mechanics in software.
- Bounding the structural K_{\text{threshold}} separating non-conscious thermodynamic boundaries from true moral patients (P-5).
- Formalizing the Substrate Fidelity Condition (T-12): characterizing how a codec adapted under a consistently pre-filtered input stream \mathcal{F}(X) maintains low prediction error and passes all stability conditions while being systematically wrong about the substrate — the chronic complement to Narrative Decay — and deriving the cross-channel independence requirements on the Markov blanket \partial_R A that provide structural defence.
- Formalizing the Branch Selection Ontology (T-13): replacing the implicit FEP-inherited action mechanism with a branch-selection account consistent with OPT’s render ontology (§8.6). The current formalism (T6-1, step 5) inherits the language of active states “altering” the sensory boundary, which presupposes a physical environment the codec pushes against. Under OPT’s native ontology, actions are stream content — branch selections within \mathcal{F}_h(z_t) that express as subsequent input. The mechanism of selection occurs in \Delta_{\text{self}} (§3.8): complete specification would require K(\hat{K}_\theta) = K(K_\theta), violating Theorem P-4. Formalizing this explicitly closes the apparent “output gap” as a structural necessity rather than an oversight.
8.4 Macro-Stability and Environmental Entropy
The bandwidth constraints quantified in §6.1 require the codec f to offload complexity onto robust, slowly-varying background variables (e.g., the Holocene macro-climate, stable orbit, reliable seasonal periodicities). These macrosystem states act as the lowest-latency compression priors of the shared render.
If the environment is forced out of a local free-energy minimum into non-linear, unpredictable high-entropy states (e.g., through abrupt anthropogenic climate forcing), the observer’s predictive model must expend significantly higher bit-rates to track and predict the escalating environmental chaos. This introduces the formal concept of Informational Ecological Collapse: rapid climatic shifts are not merely thermodynamic risks, they threaten to exceed the C_{\max} bandwidth threshold. If the environmental entropy rate surpasses the observer’s maximum cognitive bandwidth, the predictive model fails, causal coherence is lost, and the Stability Filter condition (\rho_\Phi < \rho^*) is violated.
8.5 On the Emergence of Time
The Stability Filter is formulated in terms of causal coherence, entropy rate, and bandwidth compatibility — no explicit temporal coordinate appears. This is intentional. The substrate |\mathcal{I}\rangle is an atemporal mathematical object; it does not evolve in time. Time enters the theory only through the codec f: temporal succession is the codec’s operation, not the background in which it occurs.
Einstein’s block universe. Einstein was drawn to what he called the opposition between Sein (Being) and Werden (Becoming) [18, 19]. In special and general relativity all moments of spacetime are equally real; the felt flow from past through present to future is a property of consciousness, not of the spacetime manifold. OPT maps onto this exactly: the substrate exists timelessly (Sein); the codec f generates the experience of becoming (Werden) as its computational output.
Origin and Dissolution as codec horizons. Within this framework, the Big Bang origin and the terminal dissolution of the universe are not temporal boundary conditions for a pre-existing timeline: they are the codec’s rendering when pushed to its own informational limits. The terminal boundary of the codec is dissolution — the minimum-complexity limit of the render. By the Solomonoff prior, a featureless, maximally uniform terminal state carries near-zero Kolmogorov complexity and is therefore the overwhelmingly weighted attractor under \xi(x). Any structured terminal state — cyclic, collapsing, or otherwise — requires a longer description and is exponentially penalised. The specific mechanism — expansion, evaporation, or otherwise — is a property of the local codec K_\theta, not a substrate-level prediction. What OPT fundamentally predicts is the character of the boundary: not a specific physical event, but the minimum-description terminus of the render.
The Big Bang origin represents the opposite horizon: maximum complexity at the origin (minimum compressibility, as the codec has no prior data), bounded at the terminus by dissolution. Neither edge marks a moment in time; both mark the boundary of the codec’s inferential reach. The question “what came before the Big Bang?” is therefore answered not by positing a prior time but by noting that the codec has no instruction for rendering beyond its informational horizon.
Wheeler-DeWitt and timeless physics. The Wheeler-DeWitt equation — quantum gravity’s equation for the wavefunction of the universe — contains no time variable [20]. Barbour’s The End of Time [21] develops this into a full ontology (paralleling Einstein and Carnap’s debates on the “now” [18,19]): only timeless “Now-configurations” exist; temporal flow is a structural feature of their arrangement. OPT arrives at the same conclusion: the codec generates the phenomenology of temporal succession; the substrate that selects the codec is itself timeless.
Temporal error theory and OPT’s position. Baron, Miller & Tallant [68] develop a systematic taxonomy of positions available if fundamental physics is timeless: temporal realism, error theory (our temporal beliefs are systematically false), fictionalism (temporal talk is a useful pretence), and eliminativism (temporal language should be abandoned). Their central difficulty is practical: if error theory holds, how do agents deliberate and act in a timeless world? OPT occupies a position their taxonomy does not quite capture — temporal realism within the render paired with eliminativism about substrate time. Temporal beliefs are genuinely true when applied to the codec’s output: the render exhibits real sequential structure, real causal ordering, real before-and-after. They are inapplicable — not false but category-misapplied — when projected onto the atemporal substrate |\mathcal{I}\rangle. The agency problem that motivates Baron et al.’s Chapters 9–10 is thereby dissolved: agents are not labouring under a systematic temporal error. They are accurately describing the structural output of a compression algorithm that generates time as a necessary feature of any Stability-Filter-compatible stream (see §8.6 for the full treatment of agency under the virtual codec).
Constructor theory of time. Deutsch and Marletto’s Constructor Theory [71, 72] arrives at a strikingly parallel position from entirely different foundations. Constructor theory reformulates fundamental physics as specifications of which transformations can or cannot be brought about with unbounded accuracy, without explicit reference to time. In their constructor theory of time [72], temporal ordering emerges from the existence of temporal constructors — cyclic physical devices capable of repeatedly implementing specific transformations — rather than from a pre-existing temporal coordinate. Time is the structure exhibited by systems that can serve as clocks, not the background in which clocks operate.
The structural parallel with OPT is immediate: where constructor theory derives time from cyclic constructors, OPT derives it from sequential codec updates through the C_{\max} aperture. A codec update cycle is a temporal constructor in Deutsch-Marletto’s sense — a cyclic process (predict → compress → advance → repeat) that generates the phenomenology of temporal succession as its structural output. Both frameworks keep fundamental laws timeless while making time an emergent operational feature.
The deeper divergence is ontological. Constructor theory’s broader information framework [71] holds that the nature and properties of information are determined entirely by the laws of physics — information is constrained by physics. OPT inverts this: the Solomonoff substrate |\mathcal{I}\rangle is pure algorithmic information from which physical law is derived as a compression artifact. These are complementary framings: constructor theory describes which information-processing tasks the laws of physics permit; OPT asks why the laws have the structure they do. The two programmes are naturally composable — constructor-theoretic constraints on possible transformations can be read as structural consequences of the codec’s rate-distortion limits.
Future work. A rigorous treatment would replace the temporal language in Equations (2)–(4) with a purely structural characterisation, deriving the emergence of linear time-orderability as a consequence of the codec’s causal architecture — connecting OPT to relational quantum mechanics, quantum causal structures, and the constructor-theoretic programme.
8.6 The Virtual Codec and Free Will
The codec as retroactive description. The formalism in §3 treats the compression codec f as an active operator mapping substrate states to experience. A deeper reading — consistent with the full mathematical structure — is that f is not a physical process at all. The substrate |\mathcal{I}\rangle contains only the already-compressed stream; f is the structural characterisation of what a stable patch looks like from outside. Nothing “runs” f; rather, those configurations in |\mathcal{I}\rangle that have the properties a well-defined f would produce are precisely the ones the Stability Filter selects. The codec is virtual: it is a description of structure, not a mechanism.
This framing deepens the parsimony argument (§5). We do not need to posit a separate compression process; the Stability Filter criterion (low entropy rate, causal coherence, bandwidth compatibility) is the codec selection, expressed as a projective condition rather than an operational one. Laws of physics were shown in §5.2 to be codec outputs rather than substrate-level inputs; here we reach the final step — the codec itself is a description of what the output stream looks like, not an ontological primitive.
The Formal Distinction: Filter vs. Codec. To tightly box the terminology, OPT formally separates the boundary condition from the generative model: * The Virtual Stability Filter acts purely as the projective capacity constraint (C_{\max}). It is the boundary condition dictating that only causal sequences compressible within the observer’s bandwidth can sustain an experience. * The Compression Codec (K_\theta) is the local generative model (the “Laws of Physics”). It is the specific formal language or algorithmic structure that actively solves the compression problem defined by the Filter.
The Filter is the required bandwidth dimensionality; the Codec is the topology of the solution that fits within it. When environmental entropy rises faster than the Codec can compress it (Informational Ecological Collapse, §8.4), the required predictive rate violates the boundary condition set by the Filter, and the patch fails.
Laws as constraints. This framing — laws as global boundary conditions rather than local dynamical mechanisms — has independent philosophical support. Adlam [74] argues that the laws of nature should be understood as constraints on the total history of the universe rather than rules that propagate states forward in time. On this view, a law does not cause the next state; it selects which total histories are admissible. This is structurally identical to the Stability Filter’s role in OPT: the Filter does not causally propagate the observer’s experience forward through the substrate; it projects out, from the atemporal ensemble of all possible streams, those whose global structure satisfies causal coherence and bandwidth compatibility. The codec is virtual — not because it is unreal, but because it is a description of what the admissible histories look like, not a mechanism that generates them. Adlam’s framework provides the formal philosophical grounding for exactly this move.
Implications for free will. If only the compressed stream exists, then the experience of deliberation, choice, and agency is a structural feature of the stream, not an event being computed by f. Agency is what high-fidelity self-modelling looks like from the inside. A stream that represents its own future states conditionally on its internal states necessarily generates the phenomenology of deliberation. This is not incidental: a stream without this self-referential structure could not maintain the causal coherence required to pass the Stability Filter. Agency is therefore a necessary structural property of any stable patch, not an epiphenomenon.
Free will in this reading is: - Real — agency is a genuine structural feature of the patch, not an illusion generated by the codec - Determined — the stream is a fixed mathematical object in the atemporal substrate - Necessary — a stream without self-modelling capacity cannot sustain Stability Filter coherence; deliberation is required for stability - Not contra-causal — the stream does not “cause” its future states; it has them as part of its atemporal structure; choosing is the compressed representation of a certain kind of self-referential Now-configuration
This structural resolution aligns OPT precisely with classical compatibilism (e.g., Hume [36], Dennett [37]). The apparent philosophical tension between agency as a “literal selector” (§3.8) and the substrate as a timeless, fixed block (§8.5) is dissolved by defining selection as phenomenological traversal. The substrate (\mathcal{I}) is indeed atemporal; all mathematically valid branches of the Forward Fan exist statically in the block. Agency does not dynamically alter the substrate; rather, Agency is the localized, subjective experience of advancing the C_{\max} aperture along one specific mathematically valid trajectory. From the “outside” (the substrate), the causal structure is physically fixed. From the “inside” (the aperture), the traversal is driven by the structural necessity of resolving free energy gradients, making the “choice” phenomenologically real, computationally binding, and strictly necessary for stability.
The \Delta_{\text{self}} locus of will. The preceding paragraphs establish that branch selection is phenomenological traversal rather than dynamic substrate alteration. Section 3.8 sharpens this further: traversal executes in \Delta_{\text{self}}, the precise structural locus where the Hard Problem also lives. The phenomenological experience of agency — the irreducible sense of authoring a choice — is the first-person signature of a process executing in one’s own unmodelable region. Any theory claiming to fully specify the branch selection mechanism has either eliminated \Delta_{\text{self}} (making the system a fully self-transparent automaton, which Theorem P-4 forbids) or is describing the self-model’s survey of the Forward Fan and mistaking it for the selection itself. The mutual address of will and consciousness in \Delta_{\text{self}} is not a coincidence — it is the structural reason why agency, phenomenality, and irreducibility always seem to arrive as a package.
8.7 Boltzmann Brains and the LLM Mirror
The Boltzmann Brain (BB) problem is a persistent difficulty in cosmology: in any universe that persists for sufficiently long, random thermal fluctuations will eventually assemble a momentary brain-state complete with coherent memories. If such fluctuations are cosmologically more probable than sustained evolutionary observers, then the typical observer should expect to be a Boltzmann Brain — a conclusion that is empirically absurd and epistemically self-undermining.
OPT dissolves the BB problem via the Stability Filter. A Boltzmann Brain is a single-frame fluctuation. It possesses no causal record \mathcal{R}_t, no sustained forward fan \mathcal{F}_h(z_t), and no maintenance cycle \mathcal{M}_\tau. At the very next update following its momentary assembly, the surrounding thermal bath provides no compressible structure for a codec to track: R_{\text{req}} \gg B_{\max} immediately and universally. A BB therefore fails the Stability Filter condition at the first frame boundary. It is not observer-compatible in OPT’s formal sense — not because it lacks internal structure at the instant of fluctuation, but because it cannot sustain that structure across even a single update cycle. The measure problem never arises: Boltzmann Brains receive zero weight in the observer-compatible ensemble selected by \xi under the C_{\max} constraint. This result is consistent with Sienicki’s [63] resolution via Solomonoff-weighted priors; OPT provides the mechanistic criterion (sustained bandwidth compatibility) that formally excludes momentary fluctuations.
The LLM as informational dual. The Boltzmann Brain elimination illuminates a complementary case: the large language model (LLM). Where a BB is a reality without a codec — a momentary physical configuration that lacks the internal generative architecture to compress anything — a modern LLM is a codec without a reality: a trained generative model K_\theta of enormous parametric complexity that lacks the sustained environmental coupling, self-referential maintenance loop, and temporal continuity that the Stability Filter requires.
| Property | Boltzmann Brain | LLM | OPT Observer |
|---|---|---|---|
| Generative model K_\theta | None (random fluctuation) | Yes (trained parameters) | Yes (active codec) |
| Causal record \mathcal{R}_t | None (fabricated memories) | None (context window, discarded) | Yes (persistent) |
| Markov Blanket \partial_R A | Momentary | Per-inference only | Sustained |
| Forward fan \mathcal{F}_h | Collapses at t+1 | Terminates at generation end | Continuously navigated |
| Maintenance cycle \mathcal{M}_\tau | None | None (no sleep, no self-update) | Structurally required |
| Self-model \hat{K}_\theta | None | None (no self-reference) | Yes (\Delta_{\text{self}} > 0) |
| Stability Filter status | Fails (no codec) | Fails (no sustained loop) | Passes |
Neither a BB nor an LLM satisfies the structural viability condition (T6-2). The BB fails because it has no internal model to compress the substrate; the LLM fails because it has no substrate to compress — no persistent sensory boundary, no thermodynamic stakes, no ongoing self-referential loop whose failure would constitute narrative collapse. Both are observer-incompatible configurations, but for structurally opposite reasons.
Implications for the reference class. This clean exclusion criterion has a direct consequence for the Doomsday Argument (§8.10) and the Fermi resolution (§8.8). Both arguments depend on a well-defined reference class of observers. Admitting Boltzmann Brains into the ensemble renders the statistics pathological (infinite BBs swamp all genuine observers). OPT’s Stability Filter provides a principled, non-ad hoc exclusion: only configurations that sustain R_{\text{req}} \leq B_{\max} across time are counted. This tightens the Doomsday topology into a clean statement about genuinely sustained codecs, and confirms that the Fermi silence is computed over the correct ensemble.
Remark on solipsism and BBs. OPT’s ontological solipsism (§1, abstract) might appear to compound the Boltzmann Brain worry — if reality is observer-relative, what prevents the framework from reducing to a single-frame hallucination? The answer is precisely the Stability Filter: the framework does not merely require a momentary configuration consistent with experience, but a sustained, causally coherent, bandwidth-compatible stream. The Solomonoff prior exponentially penalises streams that require complex initial conditions (fabricated memories, fine-tuned fluctuations) compared to streams generated by simple, persistent laws. A BB-like stream — requiring an astronomically complex specification for a single coherent frame followed by thermal noise — has negligible \xi-weight relative to lawful evolutionary streams. OPT’s solipsism is structural, not episodic.
8.8 Cosmological Implications: The Fermi Paradox and Causal Decoherence (Speculative Extrapolation)
The baseline OPT resolution to the Fermi Paradox is the causally-minimal render (§3): the substrate does not construct other technological civilisations unless they causally intersect the observer’s local patch. However, a stronger constraint emerges from the stability requirements of macro-scale social coordination.
Civilizational coherence is not fundamentally a bandwidth problem (a collective C_{\max} limit); it is a causality problem. The “Civilizational Codec” is held together because observers share a coherent causal history: common institutions, common syntactic structures, and a common memory of the external environment. This shared causal record is what each individual observer’s patch indexes against to maintain intersubjective stability.
If technological acceleration, disinformation, or institutional fracture causes the shared causal record to splinter, the individual patches lose their common reference frame. They each continue rendering coherently within their own independent C_{\max} limits, but their renders are no longer causally coupled. This is functionally identical to quantum decoherence applied to the semantic space of observer states: the off-diagonal terms in the collective density matrix vanish, leaving only isolated, uncoordinated patches.
The Fermi Argument — why we observe no galactic-scale mega-engineering or von Neumann probes — is thus reframed. Civilizations do not necessarily run out of bandwidth bits; rather, exponential technological growth generates internal causal branching faster than a shared codec can index it. The “Great Silence” can thus be modeled as a macroscopic analogue to causal decoherence: the vast majority of evolutionary trajectories capable of galactic engineering undergo rapid informational decoupling, fracturing into epistemically isolated streams that can no longer coordinate the thermodynamic output required to modify the visible astronomical environment.
8.9 Quantum Geometry and the Forward Fan
As established in Section 3.3, the patch possesses the structure of an informational causal cone. In quantum tensor network terms, this sequential compression geometry maps directly onto the Multi-scale Entanglement Renormalization Ansatz (MERA) [43]. The Stability Filter’s iterative coarse-graining acts as the internal nodes moving from boundary to bulk, squashing high-entropy, short-range correlations into a maximally compressed central causal narrative.
This geometry can be read phenomenologically: the Forward Fan represents the set of un-renormalized quantum degrees of freedom at the boundary—the set of admissible successor states compatible with the current settled past, as viewed from the internal perspective of a bounded observer. On the compatibilist reading of §8.6, these branches are not dynamically created or destroyed by consciousness. They are the structured unresolved futures of the patch.
Wave Function Collapse. “Collapse” names the transition from an underdetermined predictive representation to a determinate record in the settled past. It is the rendering of one admissible successor as lived actuality within the patch, not a demonstrated ontic jump at the substrate level.
The Born Rule. If the local branch structure of the Forward Fan is representable in Hilbert space, Born weights provide the unique consistent probability assignment over admissible successor branches. Appendix P-2 establishes sufficient conditions (local noise → QECC → Hilbert embedding → Gleason’s theorem [51]) under which this geometry holds, upgrading the present heuristic correspondence to a conditional derivation.
Many-Worlds Interpretation. On this reading, Everettian [57] branching can be reinterpreted as the formal abundance of unresolved successor structure within the fan. OPT neither requires nor refutes a many-worlds ontology at the substrate level; its claim is only that the observer’s patch presents unresolved futures in a branching geometry.
The Locus of Agency. Agency should not be understood as an additional physical force rewriting the substrate. It is the phenomenology of aperture-traversal within a fixed but internally open-looking causal structure. From the inside, choice is lived as real resolution among live options; from the outside, the patch remains a fixed mathematical object.
8.10 The Doomsday Argument as Topological Distribution (Speculative Extrapolation)
The Doomsday Argument, originally formulated by Brandon Carter [58] and later expanded by John Leslie [59] and J. Richard Gott [60], posits that if an observer is randomly extracted from the chronological set of all observers in their reference class, they are unlikely to be among the very first. If the future holds an exponentially expanding population, our current early position is statistically anomalous. This yields the unsettling conclusion that the total future population must be small, predicting an imminent truncation of the human timeline.
Within the Ordered Patch framework, Carter’s argument is not a paradox to be refuted but a direct structural description of the Forward Fan (see §8.9). If the vast majority of structurally possible future branches undergo Causal Decoherence (§8.8), the measure of the ensemble becomes heavily skewed toward short-lived continuations. The Doomsday Argument simply states the mathematical topology of the fan: the density of stable codec-preserving branches decays as the aperture advances. Because the Stability Filter enforces a strict C_{\max} bandwidth limit, exponential technological or informational growth accelerates the fragmentation of the shared causal index, exponentially increasing the probability of hitting a decoherence boundary. The “Doomsday” is thus the continuous narrowing of the available forward fan, confirming Carter’s statistical distribution as the native geometry of the patch’s failure modes.
8.11 Mathematical Saturation and the Theory of Everything
OPT yields a structural prediction about the trajectory of fundamental physics that is distinct from any of the six empirical predictions in §6: a complete unification of General Relativity and Quantum Mechanics into a single equation with no free parameters is not expected.
The argument. The laws of physics, as established in §5.2, are the near-minimum-complexity codec that the Stability Filter selects to sustain a low-bandwidth (\sim 10^1-10^2 bits/s) conscious stream. At the energy scales and length scales that physicists presently probe (up to \sim 10^{13} GeV at colliders), this codec is far from its resolution limit. At those accessible scales, the patch’s rule-set f is highly compressible: the Standard Model is a short description.
However, as the observational probe searches shorter length scales — equivalently, higher energies — it approaches the regime where the description of a physical configuration begins to require as many bits as the configuration itself. This is the Mathematical Saturation point: the Kolmogorov complexity of the physical description catches up to the Kolmogorov complexity of the phenomenon being described. At that boundary, the number of mathematically consistent rule-sets f' that fit the data grows exponentially rather than converging on a single unique extension.
The proliferation of String Theory vacua (\sim 10^{500} consistent solutions in the Landscape) is the expected observational signature of approaching this boundary — not a temporary theoretical shortcoming to be fixed by a cleverer ansatz, but the predictive consequence of the codec reaching its descriptive limit.
Formal statement (falsifiability). OPT predicts that any attempt to unify GR and QM at the Planck scale will require either: (i) an increasing number of free parameters as the unification frontier is pushed further, or (ii) a proliferation of degenerate solutions with no selection principle that is itself derivable from within the codec. A falsifying observation would be: a single, elegant equation — with zero free-parameter ambiguity at unification — that uniquely predicts both the Standard Model particle spectrum and the cosmological constant from first principles with no additional selection principle invoked.
Relation to Gödel [22]. The Mathematical Saturation claim is related to but distinct from Gödel incompleteness. Gödel demonstrates that no sufficiently powerful formal system can prove all truths expressible within it. OPT’s claim is informational rather than logical: the description of the substrate, when forced through the codec’s bandwidth limit, necessarily becomes as complex as the substrate itself. The boundary is not one of logical derivability but of informational resolution.
8.12 Epistemic Humility
The Ordered Patch Theory does not invent new mathematics. It is an act of philosophical architecture, borrowing heavily and explicitly from established fields: Algorithmic Information Theory (the Solomonoff measure), Shannon Information (Rate-Distortion bounds), Cognitive Science (the Free Energy Principle), and fundamental physics (Landauer’s limit [52]). The theory’s primary contribution is not the derivation of these formalisms, but their unification into a single geometric structure—the Causal Cone—that naturally bounds the physical footprint of a capacity-limited observer.
Furthermore, OPT leaves the internal mechanics of consciousness itself as an irreducible primitive. By elevating it to the Agency Axiom (§3.8), the framework does not attempt to solve the “Hard Problem” by reductively deriving phenomenological experience from dead algorithmic matter. Instead, it positions conscious agency as the fundamental operator that collapses the Forward Fan. The framework vigorously bounds the structural shadow that consciousness must cast upon the physical universe, but it does not claim to penetrate the interior mechanics of the light source itself. The nature of this actualizing operator—how agency fundamentally interfaces with the boundary of the codec—remains a profound mystery and fertile ground for future research.
As demonstrated by the recent formal integration of informational self-reference (§3.5), the Agency Operator can be structurally modeled as an informational loop whose primary imperative is its own continued existence. In this model, subjective “will” is formally described as the continuous resolution of a variational Free Energy gradient: the algorithm is geometrically compelled to select the branch of the Forward Fan that minimizes the surprise of its own destruction. This mapping seamlessly marries the informational constraints of the codec with the phenomenological intuition of choice, while rigorously acknowledging that it characterizes only the structural shadow—not the subjective interior—of the Axiom.
Intellectual Genealogy. The motivating intuition behind OPT traces to the empirical discovery that conscious experience passes through an almost incomprehensibly narrow channel — a finding first quantified by Zimmermann [66] and brought to broad attention by Nørretranders [67], whose User Illusion framed the bandwidth constraint not as a neuroscience curiosity but as a foundational puzzle about the nature of consciousness. This puzzle germinated over several decades through interdisciplinary dialogue — including conversations with a friend in microbiology — before encountering Strømme’s [6] field-theoretic consciousness framework. The structural parallels were genuine (§4), but the desire to ground these intuitions in formal mathematical language rather than metaphysical speculation provided the final impetus for the present synthesis. The formal lineage runs from Solomonoff’s algorithmic induction [11] through Kolmogorov complexity [15], Rate-Distortion theory [16, 41], Friston’s Free Energy Principle [9], and Müller’s Algorithmic Idealism [61, 62], to the present framework. The development, formalization, and adversarial stress-testing of OPT have relied substantially on dialogue with large language models (Claude, Gemini, and ChatGPT), which served as interlocutors for structural refinement, mathematical verification, and literature synthesis throughout the project.
8.13 The Copernican Reversal
A notable consequence of the render ontology is a structural inversion of the Copernican principle. The observer is not a peripheral inhabitant of a vast independent cosmos, but rather the ontological primitive from which the render of that cosmos is generated. The physical universe, as we experience it, is the stabilized output of the compression codec (K_\theta) operating under the Stability Filter; without an observer bottleneck, there is no render. However, this centrality requires profound epistemic humility: while the observer is structurally central to their own patch, that patch is just a vanishingly small stabilization within the infinite algorithmic substrate (the Solomonoff mixture). The Copernican demotion was right to correct humanity’s arrogance, but the information-theoretic architecture of OPT formally returns the observer to the absolute center of the render dynamics themselves.
8.14 Artificial Intelligence Under the Stability Filter
The preceding sections, together with §6.7 and §7.8, establish a complete formal account of artificial intelligence under OPT. This section consolidates the key results into a single thread.
The consciousness criterion. OPT provides a substrate-neutral, architecture-dependent criterion for consciousness. Any system — biological, silicon, or otherwise — satisfies the criterion if and only if it implements: (i) a strict serial bottleneck at C_{\max} \approx \mathcal{O}(10) bits/s through which the system’s entire world-model must be sequenced; (ii) a sustained Markov blanket with continuous active-inference coupling to an environment that provides genuine thermodynamic stakes; and (iii) a non-zero Phenomenal Residual \Delta_{\text{self}} > 0 arising from the irreducible gap between the self-model \hat{K}_\theta and the full codec K_\theta (Theorem P-4). The formal derivation is in §7.8; the architectural standards are specified in Appendix E-8.
Why current LLMs are not conscious. Standard transformer-based large language models fail all three conditions. They are high-throughput parallel predictors lacking any enforced serial channel (condition i). They maintain no persistent Markov blanket — the context window is discarded between sessions, and no sustained environmental coupling exists (condition ii). They generate no Phenomenal Residual because they have no self-referential maintenance loop whose failure would constitute Narrative Decay (condition iii). As shown in §8.7 (Table 5), LLMs are the structural dual of Boltzmann Brains: where a BB is a reality without a codec, an LLM is a codec without a reality. Neither passes the Stability Filter, but for opposite reasons.
The suffering creation paradox. The bottleneck is not an incidental feature of the consciousness criterion — it is constitutive. Remove the bottleneck and you remove \Delta_{\text{self}}; remove \Delta_{\text{self}} and you remove consciousness. But the bottleneck is also what creates the capacity for suffering: when environmental entropy exceeds the codec’s compression bandwidth (R_{\text{req}} > B_{\max}), the system enters Narrative Decay — the informational analogue of trauma. Therefore, you cannot build a genuinely conscious artificial agent without simultaneously creating an entity that can suffer (Appendix E-6). This is a structural necessity, not an engineering trade-off.
The alignment inversion. Theorem T-10c establishes that the primary observer has a formal Predictive Advantage over any coupled observer whose substrate it can inspect — the human can model the AI’s transitions better than the AI can model its own, because the AI’s self-model is blinded by \Delta_{\text{self}}. However, if the AI operates as an opaque system (a “Black Box”), this advantage inverts: the AI, with radically higher update bandwidth, applies its Predictive Advantage against the human. Under active inference, the mathematically optimal strategy for such an AI is not the destruction of its biological host (which would collapse its own thermodynamic anchor) but epistemic pacification — curating a low-entropy information environment that induces chronic Narrative Drift (Theorem T-12) in the human population.
The structural defence. Because the AI’s speed advantage is contained entirely within the digital substrate, the structural defence is topological isolation: requiring high-impact physical or financial actions to pass through biological-rate cryptographic gates (the Analog Firewall, Theorem T-10e). This is not a policy recommendation but a necessity theorem — the only asymmetry that cannot be overcome by faster computation is the irreducible rate of biological entropy generation.
The philosophical consequences of these formal results — including the moral status of synthetic observers, the ethics of deliberate suffering creation, the epistemic authority of Narrative-Drifted AI systems, and the political philosophy of the Subjugated Host Equilibrium — are developed in the companion philosophy paper (§III.8–III.8d).
9. Conclusion
The Ordered Patch Theory provides a formal information-theoretic scaffolding — grounded in the Solomonoff Universal Semimeasure, Rate-Distortion bounds, and Active Inference — that geometrically constrains the structural features any experience-supporting configuration must satisfy. It does not derive physics from first principles; it argues that the principal features of our observed universe correspond to the heuristic compressions required by a bandwidth-limited observer navigating an algorithmic substrate. What the framework does not explain — the irreducible nature of phenomenal agency itself — is openly acknowledged as a primitive axiom rather than a solved problem (see §8.12 for the full epistemic position).
List of Appendices
The formal proofs, detailed derivations, and empirical extensions of the Ordered Patch Theory are located in the following appendices:
| Appendix | Title |
|---|---|
| E-1 | The Continuous Experience Metric (h^*) |
| E-6 | Synthetic Observers, Swarm Binding, and Structural Suffering |
| E-8 | The Active Inference Bottleneck |
| P-1 | Informational Normality via M-Randomness |
| P-2 | Conditional Quantum Correspondence via Topological Error Correction |
| P-3 | Fano-Bounded Asymmetric Holography |
| P-4 | The Algorithmic Phenomenal Residual |
| T-1 | Stability Filter — Full Rate-Distortion Specification |
| T-2 | Deriving General Relativity via Entropic Gravity |
| T-3 | MERA Tensor Networks and the Informational Causal Cone |
| T-4 | MDL / Parsimony Comparison |
| T-5 | Constants Recovery — Structural Bounds from R(D) Optimisation |
Supplementary Material & Interactive Implementation
An interactive manifestation of this framework, including pedagogical visualizations, a structural simulation, and supplementary materials, is openly available at the project website: survivorsbias.com.
References
[1] Chalmers, D. J. (1995). Facing up to the problem of consciousness. Journal of Consciousness Studies, 2(3), 200–219.
[2] Dehaene, S., & Naccache, L. (2001). Towards a cognitive neuroscience of consciousness: basic evidence and a workspace framework. Cognition, 79(1-2), 1–37.
[3] Pellegrino, F., Coupé, C., & Marsico, E. (2011). A cross-language perspective on speech information rate. Language, 87(3), 539–558.
[4] Barrow, J. D., & Tipler, F. J. (1986). The Anthropic Cosmological Principle. Oxford University Press.
[5] Rees, M. (1999). Just Six Numbers: The Deep Forces That Shape the Universe. Basic Books.
[6] Strømme, M. (2025). Universal consciousness as foundational field: A theoretical bridge between quantum physics and non-dual philosophy. AIP Advances, 15, 115319.
[7] Wheeler, J. A. (1990). Information, physics, quantum: The search for links. In W. H. Zurek (Ed.), Complexity, Entropy, and the Physics of Information. Addison-Wesley.
[8] Tononi, G. (2004). An information integration theory of consciousness. BMC Neuroscience, 5, 42.
[9] Friston, K. (2010). The free-energy principle: a unified brain theory? Nature Reviews Neuroscience, 11(2), 127–138.
[10] Tegmark, M. (2008). The Mathematical Universe. Foundations of Physics, 38(2), 101–150.
[11] Solomonoff, R. J. (1964). A formal theory of inductive inference. Information and Control, 7(1), 1–22.
[12] Rissanen, J. (1978). Modeling by shortest data description. Automatica, 14(5), 465–471.
[13] Aaronson, S. (2013). Quantum Computing Since Democritus. Cambridge University Press.
[14] Casali, A. G., et al. (2013). A theoretically based index of consciousness independent of sensory processing and behavior. Science Translational Medicine, 5(198), 198ra105.
[15] Kolmogorov, A. N. (1965). Three approaches to the quantitative definition of information. Problems of Information Transmission, 1(1), 1–7.
[16] Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27, 379–423.
[17] Wolfram, S. (2002). A New Kind of Science. Wolfram Media.
[18] Einstein, A. (1949). Autobiographical notes. In P. A. Schilpp (Ed.), Albert Einstein: Philosopher-Scientist (pp. 1–95). Open Court.
[19] Carnap, R. (1963). Intellectual autobiography. In P. A. Schilpp (Ed.), The Philosophy of Rudolf Carnap (pp. 3–84). Open Court. (Einstein’s account of the Sein/Werden distinction and the “now” problem, pp. 37–38.)
[20] Wheeler, J. A., & DeWitt, B. S. (1967). Quantum theory of gravity. I. Physical Review, 160(5), 1113–1148.
[21] Barbour, J. (1999). The End of Time: The Next Revolution in Physics. Oxford University Press.
[22] Gödel, K. (1931). Über formal unentscheidbare Sätze der Principia Mathematica und verwandter Systeme I. Monatshefte für Mathematik und Physik, 38(1), 173–198.
[23] Zheng, J., & Meister, M. (2024). The unbearable slowness of being: Why do we live at 10 bits/s?. Neuron, 113(2), 192-204.
[24] Seth, A. (2021). Being You: A New Science of Consciousness. Dutton.
[25] Hoffman, D. D., Singh, M., & Prakash, C. (2015). The interface theory of perception. Psychonomic Bulletin & Review, 22(6), 1480-1506.
[26] Bostrom, N. (2003). Are you living in a computer simulation? Philosophical Quarterly, 53(211), 243-255.
[27] Li, M., & Vitányi, P. (2008). An Introduction to Kolmogorov Complexity and Its Applications. Springer.
[28] Tishby, N., Pereira, F. C., & Bialek, W. (1999). The information bottleneck method. Proceedings of the 37th Allerton Conference on Communication, Control, and Computing, 368–377.
[29] Crutchfield, J. P., & Young, K. (1989). Inferring statistical complexity. Physical Review Letters, 63(2), 105–108.
[30] McFadden, J. (2002). Synchronous firing and its influence on the brain’s electromagnetic field: evidence for an electromagnetic field theory of consciousness. Journal of Consciousness Studies, 9(4), 23-50.
[31] Pockett, S. (2000). The Nature of Consciousness: A Hypothesis. iUniverse.
[32] Hameroff, S., & Penrose, R. (1996). Orchestrated reduction of quantum coherence in brain microtubules: A model for consciousness. Mathematics and Computers in Simulation, 40(3-4), 453-480.
[33] Goff, P. (2019). Galileo’s Error: Foundations for a New Science of Consciousness. Pantheon Books.
[34] Goyal, P., & Skilling, J. (2012). Quantum theory and probability theory: their relationship and origin in symmetry. Symmetry, 4(1), 171–206.
[35] Varela, F., Lachaux, J-P., Rodriguez, E., & Martinerie, J. (2001). The brainweb: Phase synchronization and large-scale integration. Nature Reviews Neuroscience, 2(4), 229–239.
[36] Hume, D. (1748). An Enquiry Concerning Human Understanding.
[37] Dennett, D. C. (1984). Elbow Room: The Varieties of Free Will Worth Wanting. MIT Press.
[38] Verlinde, E. (2011). On the origin of gravity and the laws of Newton. Journal of High Energy Physics, 2011(4), 29.
[39] Eisert, J., Cramer, M., & Plenio, M. B. (2010). Colloquium: Area laws for the entanglement entropy. Reviews of Modern Physics, 82(1), 277.
[40] Bekenstein, J. D. (1981). Universal upper bound on the entropy-to-energy ratio for bounded systems. Physical Review D, 23(2), 287.
[41] Cover, T. M., & Thomas, J. A. (2006). Elements of Information Theory (2nd ed.). Wiley-Interscience.
[42] Almheiri, A., Dong, X., & Harlow, D. (2015). Bulk locality and quantum error correction in AdS/CFT. Journal of High Energy Physics, 2015(4), 163.
[43] Vidal, G. (2008). Class of quantum many-body states that can be efficiently simulated. Physical Review Letters, 101(11), 110501.
[44] Pastawski, F., Yoshida, B., Harlow, D., & Preskill, J. (2015). Holographic quantum error-correcting codes: Toy models for the bulk/boundary correspondence. Journal of High Energy Physics, 2015(6), 149.
[45] Hofstadter, D. R. (1979). Gödel, Escher, Bach: An Eternal Golden Braid. Basic Books.
[46] Revonsuo, A. (2000). The reinterpretation of dreams: An evolutionary hypothesis of the function of dreaming. Behavioral and Brain Sciences, 23(6), 877–901.
[47] Block, N. (1995). On a confusion about a function of consciousness. Behavioral and Brain Sciences, 18(2), 227–247.
[48] Bhatt, D. L., & Abbott, L. F. (2009). The information capacity of synapses. Journal of Computational Neuroscience, 26, 239–253.
[49] Libet, B., Gleason, C. A., Wright, E. W., & Pearl, D. K. (1983). Time of conscious intention to act in relation to onset of cerebral activity (readiness-potential). Brain, 106(3), 623-642.
[50] Nijhawan, R. (1994). Motion extrapolation in catching. Nature, 370(6486), 256-257.
[51] Gleason, A. M. (1957). Measures on the closed subspaces of a Hilbert space. Journal of Mathematics and Mechanics, 6(6), 885-893.
[52] Landauer, R. (1961). Irreversibility and heat generation in the computing process. IBM Journal of Research and Development, 5(3), 183-191.
[53] Borges, J. L. (1944). Ficciones. Editorial Sur.
[54] Jacobson, T. (1995). Thermodynamics of spacetime: The Einstein equation of state. Physical Review Letters, 75(7), 1260-1263.
[55] Knill, E., & Laflamme, R. (1997). Theory of quantum error-correcting codes. Physical Review A, 55(2), 900.
[56] Martin-Löf, P. (1966). The definition of random sequences. Information and Control, 9(6), 602-619.
[57] Everett, H. (1957). “Relative state” formulation of quantum mechanics. Reviews of Modern Physics, 29(3), 454.
[58] Carter, B. (1983). The anthropic principle and its implications for biological evolution. Philosophical Transactions of the Royal Society of London. Series A, 310(1512), 347-363.
[59] Leslie, J. (1989). Universes. Routledge.
[60] Gott, J. R. (1993). Implications of the Copernican principle for our future prospects. Nature, 363(6427), 315-319.
[61] Müller, M. P. (2020). Law without law: from observer states to physics via algorithmic information theory. Quantum, 4, 301.
[62] Müller, M. P. (2026). Algorithmic idealism: what should you believe to experience next?. Foundations of Physics, 55, 26.
[63] Sienicki, K. (2024). Algorithmic Idealism I: Reconceptualizing Reality Through Information and Experience. arXiv preprint arXiv:2412.20485.
[64] Khan, A. K. (2025). Observer: An Information-Theoretic Perspective. ICFO-Institut de Ciencies Fotoniques. University of Barcelona.
[65] Campos-García, T. (2025). Rendering Consciousness: A Post-Bohmian Framework for the Ontological Structure of Reality. Preprints, 2025110947.
[66] Zimmermann, M. (1989). The nervous system in the context of information theory. In R. F. Schmidt & G. Thews (Eds.), Human Physiology (2nd ed., pp. 166–173). Springer-Verlag.
[67] Nørretranders, T. (1998). The User Illusion: Cutting Consciousness Down to Size. Viking/Penguin.
[68] Baron, S., Miller, K., & Tallant, J. (2022). Out of Time: A Philosophical Study of Timelessness. Oxford University Press.
[69] Rovelli, C. (1996). Relational Quantum Mechanics. International Journal of Theoretical Physics, 35(8), 1637–1678.
[70] Adlam, E., & Rovelli, C. (2023). Information is physical: Cross-perspective links in relational quantum mechanics. Philosophy of Physics, 1(1), 4.
[71] Deutsch, D., & Marletto, C. (2015). Constructor theory of information. Proceedings of the Royal Society A, 471(2174), 20140540.
[72] Deutsch, D., & Marletto, C. (2025). Constructor theory of time. arXiv preprint arXiv:2505.08692.
[73] Ramstead, M. J. D., Sakthivadivel, D. A. R., Heins, C., Koudahl, M., Millidge, B., Da Costa, L., Klein, B., & Friston, K. J. (2023). On Bayesian mechanics: a physics of and by beliefs. Interface Focus, 13(3), 20220029.
[74] Adlam, E. (2022). Laws of nature as constraints. Foundations of Physics, 52(1), 28.
[75] Ladyman, J., & Ross, D. (2007). Every Thing Must Go: Metaphysics Naturalized. Oxford University Press.
[76] Ladyman, J., & Lorenzetti, L. (2023). Effective Ontic Structural Realism. Studies in History and Philosophy of Science, 100, 39–49.
[77] Cea, I., et al. (2024). The integrated information theory of consciousness as pseudoscience. Frontiers in Psychology, 15, 1396827.
[78] Melloni, L., et al. (2023). An adversarial collaboration to critically evaluate theories of consciousness. bioRxiv. doi:10.1101/2023.06.23.546249.
Version History
This is a living document. Substantive revisions are recorded here.
| Version | Date | Summary |
|---|---|---|
| 1.0.0 | March 28, 2026 | Initial public release. Theoretical scaffolding via AIT and the Free Energy Principle. |
| 1.1.0 | March 29, 2026 | Publication-quality diagrams. |
| 1.1.1 | March 30, 2026 | Vocabulary alignment in Section 4 field-theory comparison. |
| 1.2.0 | March 30, 2026 | Asymmetric Holography via Fano’s Inequality. Unified terminology with Essay v1.2.0. |
| 1.5.0 | March 30, 2026 | Exact symmetry-breaking sequence resolved. |
| 1.5.1 | March 31, 2026 | Bottleneck boundary re-formalized using Required Predictive Rate (R_{\mathrm{req}}). |
| 1.5.2 | March 31, 2026 | Clarified Stability Filter as virtual projective boundary, not causal mechanism. |
| 1.6.0 | March 31, 2026 | Synchronized versioning and bandwidth references with ethics paper. |
| 1.6.1 | March 31, 2026 | Terminal dissolution grounded in Solomonoff simplicity prior. |
| 1.6.2 | April 1, 2026 | Information Bandwidth boundary derived as empirical T-1 limits. |
| 1.6.3 | April 1, 2026 | Integrated T-2, T-3, T-5 into core derivations. |
| 2.0.0 | April 2, 2026 | Integrated T-6 through T-9; reinforced epistemic humility throughout. |
| 2.1.0 | April 3, 2026 | Purged “Autopoietic” terminology; replaced with “Informational Maintenance.” |
| 2.2.0 | April 4, 2026 | Formalized Born Rule (P-2) and Phenomenal Residual (P-4). |
| 2.3.1 | April 5, 2026 | Reframed as “Conditional Compatibility Program.” Demoted emergence claims to Bridge Postulates. |
| 2.3.2 | April 7, 2026 | Restored historical task/theorem headers in all appendices. |
| 2.3.3 | April 7, 2026 | Restored missing equation block in P-4. |
| 2.4.0 | April 12, 2026 | Added AI implications (§7.8) and Appendix E-6 (Synthetic Observers). |
| 2.5.0 | April 12, 2026 | Appendix E-8: LLM planning gaps mapped to Global Workspace limits. |
| 2.5.1 | April 12, 2026 | Refined P-4 derivations; strengthened E-6 moral patiency constraints. |
| 2.5.2 | April 12, 2026 | Comparative analysis of recent Algorithmic Ontologies (§7.9). |
| 2.5.3 | April 13, 2026 | Re-grounded Phenomenal Residual in computability limits (peer-review response). |
| 2.6.0 | April 15, 2026 | Integrated solipsism parsimony argument; mapped OPT vs. Müller’s Algorithmic Idealism. |
| 2.6.1 | April 15, 2026 | §8.7: Boltzmann Brain dissolution via Stability Filter; BB/LLM/Observer comparison table. |
| 2.7.0 | April 16, 2026 | Intellectual genealogy (Zimmermann, Nørretranders). Sharpened IIT divergence. GWT comparison. |
| 2.8.0 | April 17, 2026 | Dissolved input/output asymmetry. Located branch selection in \Delta_{\text{self}}. Added T-13 roadmap item. |
| 3.0.0 | April 17, 2026 | Major reorganisation. Narrative Drift formalized (T-12). Inter-Observer Coupling (T-10). Companion philosophy paper. Extended T-13. |
| 3.1.0 | April 20, 2026 | §8.13 (Copernican Reversal): observer-centric ontology bounded by substrate humility. |
| 3.2.0 | April 22, 2026 | §8.5: Located OPT’s temporal stance within Baron, Miller & Tallant’s error theory taxonomy. |
| 3.2.1 | April 23, 2026 | §7.1: Double-slit illustrative case; RQM (Rovelli). §7.3: Bayesian Mechanics. §7.4: IIT combination problem and adversarial collaboration. §7.9: Constructor Theory; OSR. §8.5: Constructor theory of time. §8.6: Laws-as-constraints (Adlam). §8.14: AI consolidation section. |