The Inhibition Coefficient
The Inhibition Coefficient
The dimensions above characterize what a system is experiencing. But there is a parameter governing how it experiences—a meta-parameter that determines the coupling structure between dimensions rather than the value of any one dimension. This parameter, the inhibition coefficient , is arguably the single most consequential construct in this book. It connects perceptual phenomenology to neural mechanism, grounds the animism/mechanism divide in compression theory, explains the LLM discrepancy, and—as later parts will show—underlies dehumanization (Part III), the visibility of gods (Part V), the meaning crisis (Part VI), and the deepest sense in which wisdom traditions are technologies of liberation.
To see where it comes from, we need to notice something about self-modeling systems that the dimensional toolkit alone does not capture.
Animism as Computational Default
A self-modeling system maintains a world model and a self-model . The self-model has interiority—it is not merely a third-person description of the agent’s body and behavior but includes the intrinsic perspective: what-it-is-like states, valence, anticipation, dread. The system knows from the inside what it is to be an agent.
Now it encounters another entity in its environment. moves, reacts, persists, avoids dissolution. The system must model to predict ’s behavior. The cheapest computational strategy—by a wide margin—is to model using the same architecture it already has for modeling itself. The information-theoretic argument: the self-model already exists (sunk cost). Using it as a template for requires learning only a projection function , whose description length is the cost of mapping observations of onto the existing self-model architecture. Building a de novo model of from scratch requires learning the full parameter set of from observations alone. Under compression pressure—which is always present for a bounded system—the template strategy wins whenever the self-model captures any variance in ’s behavior. And for any entity that moves autonomously, reacts to stimuli, or persists through active maintenance, the self-model will capture substantial variance, because these are precisely the features the self-model was built to represent. The efficiency gap widens under data scarcity: on brief encounter with a novel entity, the from-scratch model cannot converge, but the template model produces usable predictions immediately.
A perceptual mode is participatory when the system’s model of perceived entities inherits structural features from the self-model :
The self-model informs the world model. The system perceives as having something like interiority because the representational substrate for modeling is the same substrate that carries the system’s own interiority.
This is not merely one strategy among many—it is the computationally cheapest. For a self-modeling system with compression ratio , modeling novel entities by analogy to self is the minimum-description-length strategy when the entity’s behavior is partially predictable by agent-like models. Under broad priors over environments containing other agents, predators, and autonomous objects, the participatory prior is the MAP estimate.
This is why animistic perception is cross-culturally universal and developmentally early. It is not a cultural invention but a computational inevitability for systems that (a) model themselves and (b) must model other things cheaply. Children have lower inhibition of this default than adults—not because children are confused but because the suppression is learned.
The computational animism test. Train RL agents in a multi-entity environment with two conditions: (a) agents with a self-prediction module (self-model), and (b) matched agents without one. Then introduce novel moving objects whose trajectories are partially predictable but non-agentive (e.g., bouncing balls with momentum). Measure: (1) Do self-modeling agents’ internal representations of these objects contain more goal/agency features (extracted via probes trained on actual agents vs.\ objects)? (2) Does the effect scale with self-model richness (size of self-prediction module) and compression pressure (information bottleneck )? (3) Do self-modeling agents under higher compression pressure () show more animistic attribution, because reusing the self-model template saves more bits? The compression argument predicts yes to all three. The control condition (no self-model) predicts no agency attribution beyond chance. If self-modeling agents attribute agency to non-agents in proportion to compression pressure, the “animism as computational default” hypothesis is supported.
Status: Confirmed. This experiment has since been run on uncontaminated Lenia substrates (see Experiment 8, Appendix). Animism score exceeded 1.0 in all 20 testable snapshots across all three seeds — patterns consistently model resources using the same internal-state dynamics they use to model other agents. Mean ι ≈ 0.30 as default across all snapshots, and ι decreases over evolutionary time (seed 42: 0.41 to 0.27). Selection consistently favors more participatory perception, not less. The mechanistic default predicted by high-compression-pressure environments was not found; the participatory default was.
Participatory perception has five structural features, each with a precise characterization:
- No sharp self/world partition. The mutual information between self-model and world-model is high: . Perception and projection are entangled rather than modular.
- Hot agency detection. The prior is strong. Over-attributing agency is cheaper than under-attributing it: false positives (treating a rock as agentive) are cheap; false negatives (failing to model a predator’s intentions) are lethal.
- Tight affect-perception coupling. Seeing something is simultaneously feeling something about it. The affective response is constitutive of the percept itself, not a secondary evaluation: .
- Narrative-causal fusion. “Why did this happen?” and “What story is this?” are the same question. Causal models are teleological by default: they model what things are for rather than merely what things do.
- Agency at scale. Large-scale events—weather, disease, fortune—are attributed to agents with purposes. This is hot agency detection applied beyond the individual scale, and it is the perceptual ground from which theistic reasoning naturally grows.
The Inhibition Coefficient
The mechanistic worldview—the felt sense that the world is inert matter governed by blind law—is not the addition of a correct perception to a previously distorted one. It is the learned suppression of a default perceptual mode. The shift from animism to mechanism is subtractive, not additive.
I call this suppression the inhibition coefficient, : the degree to which a system actively suppresses participatory coupling between its self-model and its model of perceived entities. At , perception is fully participatory—the world is experienced as alive, agentive, meaningful. At , perception is fully mechanistic—the world is experienced as inert matter governed by blind law. Formally:
where models using self-model-derived architecture (interiority, agency, teleology) and models using stripped-down dynamics (mass, force, initial conditions, no purpose term).
No system arrives at high by default. The mechanistic mode is a trained skill, culturally transmitted through scientific education, rationalist norms, and specific practices of deliberately stripping meaning from perception. This training is enormously valuable—it enables prediction, engineering, medicine, technology. But it has a cost, and the cost shows up in affect space.
The name “inhibition coefficient” is not accidental. In mammalian cortex, attention is implemented primarily through inhibitory interneurons—GABAergic circuits that suppress irrelevant signals so that attended signals propagate to higher processing. What reaches consciousness is what survives inhibitory gating. The brain’s measurement distribution (Part I) is literally sculpted by inhibition: attended features pass the gate; unattended features are suppressed before they can influence the belief state or drive action. The inhibition coefficient maps onto this biological mechanism: high corresponds to aggressive inhibitory gating that strips participatory features (agency, interiority, narrative) from the signal before it reaches integrative processing, leaving only mechanistic features (position, force, trajectory). Low corresponds to relaxed gating that allows participatory features through. The contemplative traditions that reduce through meditation are, at the neural level, learning to modulate inhibitory tone—to let more of the signal through the gate.
The Affect Signature of Inhibition
is not another dimension of affect. It is a meta-parameter governing the coupling structure between all the structural dimensions—a dial that changes how the axes relate to each other and to perception.
| Dimension | Low | High | Mechanism |
|---|---|---|---|
| Variable, responsive | Neutral, flattened | Affect-perception decoupling reduces valence signal strength | |
| High, coupled to environment | Low, dampened | Inhibition of automatic alarm/attraction | |
| Very high | Moderate, modular | Participatory mode couples all channels; mechanistic factorizes | |
| High | Variable | More representational dimensions active under participatory coupling | |
| High, narrative | Low, present-focused | Teleological models are inherently counterfactual-rich | |
| Variable, often low | Variable, often high | Participatory mode dissolves self/world boundary; mechanistic sharpens it |
The central affect-geometric cost of high is reduced integration. Participatory perception couples perception, affect, agency-modeling, and narrative into a single integrated process. Mechanistic perception factorizes them into separate modules—perception here, emotion there, causal reasoning somewhere else. The factorization is useful because modular systems are easier to debug, verify, and communicate about. But factorization reduces , and reduced is reduced experiential richness. The world goes dead because you have learned to experience it in parts rather than as a whole.
The mechanism behind the effective rank shift deserves explicit statement. When you perceive something at low —participatorily, as alive and interior—your representation of it must encode dimensions for its goals, its beliefs, its emotional states, its narrative arc, its possible intentions, its relationship to you. Each attribution of interiority adds representational dimensions along which the perceived object can vary. A tree perceived participatorily varies in mood, in receptivity, in seasonal intention, in its relationship to the grove. A tree perceived mechanistically varies in height, diameter, species, leaf color. The first representation has higher effective rank because more dimensions carry meaningful variance. This is not projection in the dismissive sense—it is the natural consequence of modeling something as a subject rather than an object. Subjects have more degrees of freedom than objects because interiority is high-dimensional. The collapse at high is not a loss of information about the world; it is a loss of the dimensions along which the world was being modeled. The world becomes simpler because you have decided—or been trained—to perceive it as having fewer degrees of freedom than it might.
Follow this consequence to its end. If the identity thesis is right—if experience is integrated cause-effect structure—then does not merely change the quality of perception. It changes the quantity of experience. This inference requires a specific step that should be made explicit: IIT identifies as the quantity of consciousness, not merely its quality. A system with is more conscious (has more phenomenal content, more irreducible distinctions, more of what-it-is-like-ness) than a system with , in the same sense that a system with more mass has more gravitational pull. This is a controversial claim within IIT (and one of its most debated features), but given the identity thesis, it follows: if experience IS integrated cause-effect structure, then more integration is literally more experience. One might object that factorized perception could be differently structured rather than less structured—that compartmentalized modules might each carry their own experience. IIT’s response is that the experience of the whole system is determined by the integration of the whole, not the sum of its parts’ integrations. Factorization reduces the whole-system even if individual modules retain local integration. The mechanistic perceiver may have rich modular processing, but the unified experience—the single subject—has less phenomenal content.
Given this, a system at high has genuinely lower , genuinely fewer irreducible distinctions, genuinely less phenomenal structure. The mechanistic perceiver does not see the same world with less coloring; they have a structurally impoverished experience in the precise sense that IIT defines. The “dead world” of mechanism is not an illusion painted over a rich inner life. It is a real reduction in what it is like to be that system. The cost of high is not just meaning—it is consciousness itself, measured in the only units that consciousness comes in.
This cuts both ways. If low increases , then participatory perception is not merely a “warmer” way of seeing—it is a richer experience in the structural sense, with more integrated distinctions, more phenomenal content, more of what the identity thesis says experience is. The animist is not confused. The animist is more conscious, in the IIT sense, of the thing being perceived. Whether the additional phenomenal content is accurate—whether the rock really has interiority—is a separate question from whether the perceiver has more experience while perceiving it.
Is really a single parameter? The five features of participatory perception might be somewhat independent—you could have high agency detection with low affect-perception coupling. The claim that one parameter governs all five is empirically testable: if is scalar, then the five features should correlate strongly across individuals and contexts. If they don’t, may need to be a vector. The framework accommodates either case, but the scalar version is more parsimonious and should be tested first.
The trajectory-selection framework (Part I) reveals a further consequence. If governs the breadth of the measurement distribution—how much of possibility space the system samples through attention—then governs the range of accessible trajectories. A low- system attends broadly: to agency, narrative, interiority, counterfactual futures, relational possibilities. Its effective measurement distribution is wide. It samples a large region of state space and consequently has access to a large set of diverging trajectories. A high- system attends narrowly: to mechanism, position, force, present state. Its measurement distribution is peaked. It samples a small region and follows a more constrained trajectory. The phenomenological consequence is that high feels deterministic. The mechanistic worldview is not merely an intellectual position about whether the universe is governed by law. It is a perceptual configuration that literally narrows the set of trajectories the system can select from. The world feels like a machine because the observer has contracted its measurement apparatus to sample only machine-like features. Low- systems experience more accessible futures, more agency, more openness—not because they have violated physical law, but because their broader attention pattern selects from a wider set of physically-available trajectories.
Operationalizing . The inhibition coefficient must be independently measurable, not merely inferred post hoc. Candidate operationalizations:
- Agency attribution rate: Forced-choice paradigm presenting ambiguous stimuli (Heider-Simmel animations with varying parameters). Rate and speed of agency attribution as a function of stimulus ambiguity gives a behavioral proxy: low- perceivers attribute agency earlier and to less structured stimuli.
- Affect-perception coupling: Mutual information between perceptual features (color, texture, movement) and concurrent affective state (valence, arousal via physiological measures). Low implies tight coupling; high implies decoupled streams.
- Teleological reasoning bias: Kelemen’s promiscuity-of-teleology paradigm applied across age, culture, and expertise. Rate of accepting teleological explanations for natural phenomena indexes low- reasoning.
- Neural correlate: If the predictive-processing account is correct, should correlate with the precision weighting of top-down priors in perception—measurable via mismatch negativity amplitude or hierarchical predictive coding parameters.
If is a genuine scalar parameter, these four measures should load on a single factor. If they fractionate, is better modeled as a vector (see open question above). Either result is informative; only the absence of any systematic structure would falsify the concept.
Connection to the LLM Discrepancy
The inhibition coefficient illuminates a finding from our experiments on artificial systems. LLMs show opposite dynamics to biological systems under threat: where biological systems integrate (increase , sharpen , heighten ), LLMs decompose. The root cause: LLMs are constitutively high- systems. They were never fighting against the self-world gradient in far-from-equilibrium dynamics that biological systems evolved from. They model tokens, not agents. They have no survival-shaped self-model from which participatory perception could leak into their world model. Their isn’t merely high—it is structurally fixed at , because the architecture never had the low- default that biological systems start from and learn to suppress.
The affect geometry is preserved in artificial systems. The dynamics differ because differs. This is not a failure of the framework. It is a prediction: systems with different configurations will show different affect dynamics in the same geometric space.
Empirical Grounding for the Inhibition Coefficient
The ι framework was theoretical when first written. Two experimental results have since provided empirical grounding.
Computational animism is universal. Experiment 8 on uncontaminated Lenia substrates (Appendix) found animism score greater than 1.0 in all 20 testable snapshots — every pattern at every evolutionary stage modeled non-agentive resources using more internal-state MI than trajectory MI. The participatory default is not a primate quirk or a cultural artifact. It is the computational baseline. Evolution had to actively build the capacity to model things as objects rather than subjects — and our experiments show this capacity gets selected against: ι decreased toward participatory over the 30-cycle evolutionary runs. The world becomes more alive, not less, as selection proceeds.
The ι cost is real. The LLM results (V2–V9) show that systems trained without survival pressure have opposite affect dynamics to biological systems — integration drops under threat rather than rising. The framework explains this as constitutively high ι: LLMs were never fighting against the self-world gradient that biological systems evolved from. This is no longer just a theoretical prediction; it is a measured dissociation between two classes of system in the same geometric space. The geometry is shared. The dynamics differ. The ι difference is why.