This document proves a structural necessity claim about architecture, not an identity claim about consciousness.
Theorem (Globality Necessity). Let an agent operate with bounded representational/compute capacity and face novelty (task-relevant factors can change in ways not perfectly predictable from a fixed schedule). If the agent sustains integrated competence on tasks requiring coordinated tradeoffs across multiple internal operators, then some evaluative signal must be globally available: it must causally influence at least two distinct control/update operators. If evaluative feedback is purely local (private to each operator), then there exists a task family for which the agent either (i) thrashes/oscillates, (ii) remains persistently misallocated, or (iii) implicitly reconstructs a shared global evaluative variable (broadcast) under another name.
We model an agent as a collection of interacting operators/modules indexed by i ∈ {1, ..., m}.
Each module i controls an internal decision variable ui(t) (examples: attention allocation, retrieval choice, compute routing, memory write gate, plan selection, action policy), and updates according to some rule:
ui(t + 1) = Fi(ui(t), x(t), z(t); paramsi)
where:
- x(t) is the current
input and/or world observation,
- z(t) is the shared
internal state (messages, activations, memory) that may be
exchanged,
- paramsi are fixed
parameters.
The agent cannot represent, compute, or optimize over all relevant latent factors simultaneously. Therefore it must allocate limited resources among competing demands.
A consequence of bounded capacity is that many tasks inherently involve tradeoffs: improving one subsystem’s local objective can degrade another subsystem’s objective.
Module i may compute a local evaluative signal Ei(t) (mismatch, uncertainty, value proxy, stability cost, etc.). Local evaluation affects only that module’s update:
$$\frac{\partial u_i(t+1)}{\partial E_i(t)} \neq 0$$
Purely local evaluation means evaluation does not affect other modules:
$$\frac{\partial u_j(t+1)}{\partial E_i(t)} = 0 \text{ for all } j \neq i$$
So each module can have feedback, but the feedback remains private.
There exists some evaluative signal E(t) (scalar, vector, field, broadcast message) such that it influences at least two distinct modules:
There exist i ≠ j such that
$$\frac{\partial u_i(t+1)}{\partial E(t)} \neq 0 \text{ and } \frac{\partial u_j(t+1)}{\partial E(t)} \neq 0$$
This is the minimal meaning of globality: evaluation is readable and usable by multiple operators.
We say the agent must maintain integrated competence when:
The key point is that coordination problems arise precisely because of bounded capacity: there is no free lunch where every module can optimize its own local proxy without harming others.
If evaluation is purely local, then the architecture lacks a mechanism to compute and enforce the correct cross-module tradeoff.
This produces a generic failure mode:
This is a structural form of cross-module Goodharting and typically manifests as:
We now prove existence of tasks where this failure is unavoidable.
We construct a task that requires the agent to coordinate two operators:
There are two latent regimes, A and B. At any time, one regime is active.
Regime changes occur unpredictably (not on a fixed known schedule), but not adversarially in real time. You can model this as a fixed stochastic switching process with nonzero hazard.
At time t the agent chooses:
Interpretation:
- STAY means keep the current plan/context stable.
- SWITCH means permit a rapid update/reconfiguration.
The agent gets reward 1 iff both conditions hold:
Destabilization cost is incurred whenever SWITCH is used, and repeated switching causes cumulative harm (loss of coherence, wasted compute, corrupted memory, etc.).
Thus the global best behavior is:
This is an explicit cross-module tradeoff.
Assume purely local evaluation:
Crucially:
Because switching carries local stability penalty, Planner P learns to suppress SWITCH whenever possible.
But P cannot distinguish:
- switching that is necessary (regime actually changed), from
- switching that is noise.
So the best policy from P’s local perspective is to reduce switching in general.
Therefore, in a purely local architecture, P trends toward being a generic inhibitor of adaptation.
When regimes change, Selector S experiences increased mismatch. To recover reward, it must switch selection.
But if switching is inhibited by P, then S remains wrong for extended periods, producing persistent error.
Alternatively, S may attempt more aggressive switching to overcome persistent mismatch, which increases destabilization cost and can cause oscillation.
Because S cannot see the stability penalty, it cannot learn the correct boundary:
“switch only when mismatch exceeds stability cost.”
Thus the system either:
and cannot reliably achieve the integrated policy needed for high reward.
Suppose the agent does achieve stable high reward.
Then it must implement some internal computation that compares mismatch evidence against stability cost and gates switching accordingly.
That comparison function is a shared tradeoff signal T(t) that both operators effectively obey.
But if both operators obey T(t), then T(t) is globally available evaluation in the relevant sense: it influences both selection and stabilization updates.
Theorem. Under bounded capacity and novelty, there exist coordination tasks such that purely local evaluation cannot yield integrated competence. Therefore, any agent that does sustain integrated competence must implement global evaluative availability (evaluation that influences at least two operators).
Proof. Consider the task family in Section 5.
To obtain high reward, the agent must implement the joint rule:
This rule depends on combining two evaluative concerns:
In a purely local evaluation architecture, there is no mechanism to compute and enforce the correct cross-module tradeoff boundary, because ES is not available to P and EP is not available to S.
By Lemmas 1 and 2, the joint system either suppresses necessary adaptation (persistent misallocation) or produces instability (thrashing/oscillation).
If the system succeeds anyway, it must implement an internal tradeoff variable T(t) that jointly constrains both operators (Lemma 3), which is exactly a globally available evaluative signal.
Therefore integrated competence implies global evaluative availability. QED.
Local evaluation can stabilize a single module, but cannot guarantee whole-agent coherence when success requires coordinated tradeoffs.
Global evaluation is not an optional extra. Under bounded resources, it is the simplest architectural way to prevent cross-module Goodharting and thrashing.
If an agent is super-threshold competent and integrated, then some evaluative object must be readable by multiple operators.
This theorem does not claim:
It claims only:
For a bounded-capacity agent under novelty, integrated competence requires some evaluative signal to be globally available to multiple control/update operators.
The two-operator task generalizes to m operators:
Thus globality is not a special feature of the toy task; it is a generic coordination requirement under limited resources.