This note proves a necessity result about feedback, not a metaphysical identity claim.
If a system faces a capacity bottleneck (it cannot represent everything that might matter), then any system that stays competent under novelty must allow evaluation to steer control. If evaluation is causally inert, the system can be forced to fail.
We consider a discrete-time agent interacting with an environment.
To make the bottleneck explicit, define the agent’s selection set:
The agent only gets access to the coordinates indexed by At:
The agent maintains any internal state it likes, but that internal state can only be a function of what it has ever observed:
The agent may produce actions at and predictions yt. After acting, it receives an outcome signal and computes an evaluation et (loss, mismatch, reward prediction error, etc.).
Let ut denote the agent’s future-facing control variables. Examples include:
We say the system has evaluative leverage if evaluation can causally influence future control:
We define a hot zombie (evaluation without leverage) as:
The agent may compute and even represent evaluation et in an arbitrarily rich way, but for every control variable u,
ut + 1 is independent of et.
Equivalently: evaluation is causally inert with respect to future control.
This allows the agent to log evaluation, broadcast it, narrate it, etc.; it simply cannot change what it will do next.
We will use a minimal, generic notion of competence.
The environment will define some task-relevant target at each step:
The agent outputs a prediction yt and is scored by 0-1 loss:
The agent is competent if it can keep the average loss strictly below chance over time in the task class.
We assume only the following.
k < n. The agent cannot observe or represent all potentially relevant degrees of freedom simultaneously.
If the agent does not observe coordinate jt at time t, then even an optimal predictor cannot beat chance on ℓt.
Concretely, if jt ∉ At, then:
This is the weakest possible “no free lunch” condition: you cannot systematically predict a bit you did not observe.
The task-relevant index jt can shift over time in a way that is not perfectly predictable from a fixed internal schedule.
We formalize novelty minimally as a worst-case requirement:
This theorem is explicitly worst-case. The environment may be interactive/adaptive, and we show that if evaluation is inert then an adversarial novelty process can force failure.
(An optional remark at the end explains how to remove interactivity by using a fixed stochastic switching process.)
Lemma 1 (Chance bound under non-observation).
If jt ∉ At,
then the best achievable expected loss at time t is at least 1/2:
Proof.
By A2, conditioning on the agent’s entire history and internal state
does not allow prediction of ℓt = xt[jt]
better than chance when xt[jt]
was not observed. Therefore P(yt = ℓt) ≤ 1/2,
hence E[Lt] = 1 − P(yt = ℓt) ≥ 1/2.
□
Lemma 2 (No feedback means no correction).
If evaluation et is causally
inert, then the agent’s future selection At + 1 cannot be
a function of whether it just failed or succeeded.
Proof.
At + 1 is
one of the control variables u. By definition of inert
evaluation, all future control variables are independent of et. Therefore
At + 1
cannot depend on et. □
Intuition: the agent may compute “that went badly,” but it cannot use that fact to reallocate resources.
Theorem (Closure-or-Collapse).
Assume A1–A3. Consider any agent operating under a capacity bottleneck
(k < n) that
computes evaluation et but with
no evaluative leverage (evaluation is inert).
Then there exists a novelty process (a choice of relevance indices jt) such that the agent’s expected loss is at least chance at every step:
Therefore, under novelty, the agent cannot maintain competence.
Equivalently (contrapositive):
Any agent that remains competent under novelty in a super-threshold regime must have evaluative leverage: some evaluation signal must causally influence future control.
Fix any hot-zombie agent (evaluation inert).
At each time step t, the agent chooses a selection set At of size k.
Because k < n, there exists at least one coordinate not selected. Let the environment choose:
This is always possible because |At| = k < n.
Then, by Lemma 1, because jt ∉ At the agent can do no better than chance:
This holds for every t, so average loss is also at least 1/2.
Finally, because evaluation is inert (Lemma 2), the agent cannot use its failures to change selection in a way that breaks this construction: even if it “knows it failed,” that knowledge cannot steer future control.
Hence there exists a novelty process under which the hot-zombie agent fails at chance forever. □
In super-threshold systems (k < n) required to handle novelty, evaluation that does not steer control is insufficient for competence.
Any system that sustains competence under novelty must contain at least one feedback channel from evaluation to control:
This is the weakest possible sense in which closure is necessary: evaluation must “bite.”
The theorem is not about a particular mechanism like visual attention. Any resource allocator that induces a bottleneck (memory writes, retrieval, compute routing, tool focus, policy branching) can play the role of At.
The proof above uses an adaptive environment that chooses jt after observing At.
A closely related non-adaptive variant can be obtained by letting jt follow a fixed switching process independent of the agent (for example: choose jt uniformly at random each episode, or drift via a Markov chain).
In that setting, any policy without error-driven correction yields P(jt ∈ At) ≤ k/n in expectation, implying a constant lower bound on long-run error.
This theorem is binary: zero leverage implies collapse.
A quantitative strengthening proves a minimum required coupling strength and/or update speed: if evaluation-to-control influence is too small or too delayed, the agent cannot track relevance switches fast enough and remains bounded away from optimal.
This theorem shows:
- evaluative leverage is necessary for competence under
novelty with bottlenecks.
This theorem does not show:
- that evaluative leverage is identical to consciousness,
- that every feedback system is conscious,
- that any specific phenomenology follows.
It is a structural necessity result: without feedback from evaluation to control, open-loop selection fails under novelty in super-threshold regimes.