Most people reach for the word “agency” when they mean the ability to act, to make choices, to do things in the world. But this folk concept is too broad to do any real theoretical work. A thermostat “acts”—it turns on the heater. A river “acts”—it carves a canyon over millennia. A chess engine “acts”—it moves pieces with superhuman precision. None of these are agents in any meaningful sense, and yet all of them satisfy the naive definition.
The mechanistic account says: agency is just complex mechanism. Inputs → processing → outputs. Everything is mechanism all the way down, so agency is an illusion projected onto sufficiently complex systems. The cognitive science version of this is computationalism—the mind is a computer, agency is program execution, and the feeling of choosing is epiphenomenal froth on deterministic machinery. This view has the virtue of parsimony and the vice of explaining away the very thing it set out to explain.
The functionalist account says: if it walks like an agent and talks like an agent, it is an agent. Internal organization does not matter, only input-output behavior. This is the Turing test intuition extended to all of agency. A thermostat is a very simple agent; a human is a very complex one; the difference is degree, not kind. But functionalism makes agency too cheap. By its lights, a lookup table with the right entries is an agent, and this seems wrong in a way that is hard to articulate but impossible to ignore.
Both accounts miss something crucial. Barandiaran, Di Paolo, and Rohde (2009) argued that genuine agency requires three specific conditions, all present simultaneously. Not one, not two—all three. Without any one of them, you do not have an agent. You have something else—something that may be complex, responsive, even impressive, but not genuinely agentive. The three conditions are: individuality, interactional asymmetry, and normativity.
Click the cards below to expand common mistaken accounts of agency and see why each falls short:
To make the three conditions concrete, consider how they apply to a range of familiar systems. Each system below is a candidate for agency. Most fail.
| System | Individuality? | Asymmetry? | Normativity? | Agent? |
|---|---|---|---|---|
| Thermostat | No — externally assembled | No — fixed coupling | No — nothing at stake | No |
| Chess engine | No — no self-maintenance | Partial — selects moves | No — winning is externally defined | No |
| Candle flame | Yes — self-maintaining | No — cannot modulate coupling | Yes — precarious | No |
| E. coli | Yes — autopoietic | Yes — tumble/run modulation | Yes — must eat or die | Yes |
| Roomba vacuum | No — externally designed | Partial — navigates obstacles | No — nothing at stake | No |
| Self-driving car | No — externally constituted | Yes — modulates many couplings | No — “safety” norms are designed in | No |
| Human being | Yes — immune system, metabolism | Yes — continuous sensorimotor modulation | Yes — material precariousness | Yes |
The table makes visible a pattern: technological systems tend to have interactional asymmetry (they modulate couplings with their environments) but lack individuality and normativity. Living systems tend to have all three. The gap is not about complexity or capability but about organizational type.
The agent must be a distinguishable entity with a self-generated and self-maintained identity. Not just any bounded thing—a rock has boundaries but no individuality in this sense. A cloud has a recognizable form but no self-generated identity. The boundaries must be produced and maintained by the system’s own ongoing activity.
This connects directly to autopoiesis (Part IV) and organizational closure (Parts I–III): the system must be operationally closed, producing the processes that produce it. Its identity is not imposed from outside but emerges from its own dynamics. The cell membrane is not placed around the cell by an external agent—the cell produces the membrane that contains the reactions that produce the membrane. This circularity is the engine of individuality.
Individuality is not about having a body or occupying space. It is about having an organization that distinguishes self from non-self through the system’s own ongoing activity. The membrane of a cell, the immune system of an organism, the behavioral coherence of an animal—all are different expressions of self-generated individuality at different scales. What they share is this: the boundary between system and environment is not merely structural but processual. It is constantly being produced, maintained, and repaired by the system’s own activity.
Consider the contrast. A crystal has a boundary, a lattice structure, a definite form. But its form is imposed by the laws of molecular packing, not generated by the crystal’s own activity. The crystal does not do anything to maintain its crystallinity. A cell, by contrast, is constantly working to maintain its membrane integrity, to import nutrients, to export waste, to repair damage. Stop the work and the cell disintegrates. The boundary persists only because the system actively sustains it.
There is a spectrum of individuality that is worth mapping. At the lowest end, a dissipative structure like a Bénard convection cell maintains a recognizable pattern through continuous energy flow, but its identity is imposed by boundary conditions — the temperature difference between the heated plate and the cooled surface. Remove the temperature gradient and the cells vanish instantly; they were never self-individuated. At a higher level, a flame maintains itself through a self-sustaining chemical reaction, but cannot repair damage to its organization or adapt to perturbations. At a still higher level, a living cell not only maintains itself but actively repairs damage, adjusts to changing conditions, and produces the very components that constitute its boundary. This last case is genuine individuality in the strong sense.
The question for artificial systems is whether computational processes can achieve genuine individuality. A software process that monitors its own health, restarts failed components, updates its own parameters, and maintains a persistent identity across time has something resembling individuality. But is the resemblance deep or superficial? The process runs on hardware it did not produce, in an operating system it did not create, powered by electricity it has no role in generating. The closure is partial at best. Whether partial closure is enough for genuine individuality, or whether full material closure is required, remains contested.
The following table contrasts individuality across different types of systems. Not all boundaries are equal, and not all identities are self-generated:
The agent must modulate its coupling with the environment. It does not just receive perturbations passively—it actively shapes the interaction. The causal flow is asymmetric: the agent influences how the environment affects it, not just vice versa.
This is what distinguishes agency from mere mechanism. A thermostat responds to temperature, but it does not modulate how it couples to temperature. The threshold is fixed, the response is fixed, the relation is fixed by the engineer who built it. E. coli does something fundamentally different—by switching between tumbling and running, it changes its relation to the chemical gradient. It does not move the food, but it changes how it relates to the food. The bacterium is the source of the modulation, not the medium.
The asymmetry is subtle and easy to mischaracterize. Both agent and environment affect each other—the environment acts on the agent, and the agent acts on the environment. But the agent additionally regulates the coupling parameters themselves. It is not just inside the interaction; it is modulating the interaction. The difference is between being a variable in an equation and being able to change the equation’s coefficients.
A ball rolling down a hill is coupled to the hill—gravity acts on the ball, friction acts between ball and surface. But the ball does not modulate any of these parameters. It is a passive participant in a fully determined interaction. An animal walking down a hill is also coupled to the hill by gravity and friction, but it additionally modulates its gait, shifts its weight, chooses where to place its feet. The animal is the source of modulation in the coupling. The hill is not.
Interactional asymmetry is about the agent being the source of modulation in the agent-environment coupling, not about the agent being more powerful than the environment. A bacterium is vanishingly small compared to its medium, but it is the source of its own behavioral transitions. The ocean does not decide when the bacterium tumbles. The bacterium decides—or rather, its own internal dynamics determine the switching rate, and that is enough.
The concept of interactional asymmetry also clarifies a common confusion about agency and control. An agent does not need to control its environment. It does not need to dominate, direct, or overpower the external world. It needs only to modulate the parameters of its own coupling. A bacterium in a turbulent ocean controls nothing about the ocean. But it controls the one thing that matters: the probability of its own tumbling. This minimal modulation — adjusting a single parameter in response to a temporal gradient — is sufficient for genuine interactional asymmetry.
This has consequences for how we think about power and agency. Agency is not proportional to power. A whale is more powerful than a bacterium, but both are equally agents. The difference is in the complexity and scope of their coupling modulation, not in the presence of modulation. A human with locked-in syndrome, unable to move anything but their eyes, is still an agent — they modulate their eye movements in relation to their environment. Agency can be diminished, constrained, impoverished, but it persists as long as the three conditions hold.
The agent’s activity must matter to it—there must be norms according to which the agent’s interactions can go better or worse, and these norms must be generated by the agent’s own organization, not imposed from outside.
This is where adaptivity (Part V) becomes crucial. Autopoiesis alone gives you alive/dead—a binary, no gradients, no stakes beyond the single threshold of dissolution. Adaptivity gives you the viability set with a graded interior. Now the agent can be closer to or further from its viability boundary. Now there is a sense in which some states are better than others for the agent, not according to some external criterion but according to the agent’s own constitutive organization.
The normativity is intrinsic—it comes from the system’s own self-maintaining organization, not from an external designer or evaluator. This is why drive systems in AI are not genuine normativity: the drives were designed, the decay rates were set by someone else, the satisfaction conditions were externally specified. The bacterium’s norms arise from the fact that it must eat or it ceases to exist. Nothing has to tell it that starvation is bad. The badness of starvation is constitutive—it is identical with the approach toward the viability boundary.
This is a radical claim. It says that values, in the most minimal sense, are not added to a living system from outside. They are generated by the system’s own precariousness. The fact that the system can cease to exist, and that some states are closer to cessation than others, is sufficient to ground a normative dimension. The system does not need to “know” that it is precarious in any cognitive sense. The precariousness is enough.
Three consequences follow from grounding normativity in precariousness:
The distinction between intrinsic and extrinsic normativity is sharp and consequential. An intrinsic norm is one generated by the system’s own constitutive organization — the bacterium needs glucose because its metabolism requires it. An extrinsic norm is one imposed from outside — the robot “needs” to charge its battery because the engineer programmed a low-battery warning. The behaviors may look identical from the outside, but the organizational status is fundamentally different.
Why does this distinction matter? Because it determines whether the system is a genuine center of concern or merely a system that behaves as if it had concerns. A system with intrinsic normativity has a perspective — a point of view from which things can go well or badly. A system with only extrinsic normativity has no such perspective; it merely executes behaviors that an external observer interprets as concern-driven. This is the difference between an agent and a sophisticated puppet, and it turns entirely on the source of the norms.
All three conditions must be present simultaneously. Remove any one, and you get something recognizable but categorically different from genuine agency. The intersection is not a continuum—it is a conjunction. Each condition is individually necessary and only jointly sufficient.
The interactive diagram below maps the logical space. Hover or click on each region to see what occupies it. The full intersection—where all three conditions meet—is the domain of genuine agency.
The table below summarizes what each partial intersection corresponds to in concrete systems:
| Conditions Present | Missing | Example System |
|---|---|---|
| All three | — | E. coli, animals, organisms generally |
| Individuality + Asymmetry | Normativity | Self-repairing autonomous robot (nothing at stake) |
| Individuality + Normativity | Asymmetry | Candle flame (precarious, self-maintaining, passive) |
| Asymmetry + Normativity | Individuality | Designed AI with drives (externally constituted identity) |
| Individuality only | Asymmetry, Normativity | Autopoietic system at equilibrium |
| Asymmetry only | Individuality, Normativity | Adaptive control system |
| Normativity only | Individuality, Asymmetry | Far-from-equilibrium dissipative structure |
| None | All three | Rock, puddle, dead mechanism |
Notice that the logical space is not merely academic. Each region corresponds to real systems that we encounter and sometimes mistakenly call “agents.” The Venn diagram is a diagnostic tool. When someone claims a system has agency, you can ask: does it have self-generated individuality, or was its identity externally imposed? Does it modulate its own coupling parameters, or does it merely respond within fixed couplings? Do its interactions matter to it because of its own constitutive organization, or are the “stakes” imposed by an external designer?
The three conditions are not a checklist to be ticked off mechanically. They are deeply intertwined. Individuality is what makes the agent a something that can be the source of modulation. Interactional asymmetry is how individuality expresses itself in the world. Normativity is what makes the modulation matter. Remove any one leg and the structure collapses into something categorically different.
It is worth noting what the framework does not claim. It does not claim that the three conditions are easy to assess in borderline cases. It does not claim that there are no interesting systems in the partial-overlap regions. And it does not claim that agency is all-or-nothing. The conditions can be met more or less fully, and there is room for degrees: a system with stronger individuality, more pervasive interactional asymmetry, and deeper normativity is more agentive than one with weaker versions of the same. But the floor — the minimum threshold — requires all three present in at least some minimal degree.
This is important because it resists two common temptations. The first is agency inflation — attributing agency to any system that acts or responds, from thermostats to chatbots. The second is agency deflation — denying agency to anything short of full human consciousness. The three-conditions framework navigates between these extremes. It is inclusive enough to recognize E. coli as a genuine agent, and exclusive enough to deny agency to a chess engine. The boundary it draws is principled, not arbitrary.
The paradigm case. Escherichia coli chemotaxis is the simplest known system that satisfies all three conditions simultaneously. It is, in a precise sense, the floor of agency—the minimal system that earns the designation.
The molecular mechanism is remarkably well understood. The chemotaxis signaling network consists of methyl-accepting chemotaxis proteins (MCPs) that detect attractants and repellents, a histidine kinase (CheA) that phosphorylates the response regulator CheY, and an adaptation system (CheR and CheB) that provides the crucial temporal comparison. When CheY is phosphorylated, it binds to the flagellar motor and increases the probability of clockwise rotation, which causes tumbling. When attractant concentration is increasing, CheA activity decreases, less CheY is phosphorylated, and the bacterium runs longer.
The adaptation system is the key to temporality. CheR continuously methylates the MCPs, gradually restoring their signaling activity after an attractant stimulus. This methylation acts as a molecular memory — it records the recent past concentration, allowing the system to compare “now” to “a moment ago.” The adaptation timescale is about one second, which defines the temporal depth of the bacterium’s present. This is minimal agency: a half-second memory implemented in protein methylation states.
Below: a chemotaxis simulator. The heat map shows nutrient concentration (click anywhere to move the nutrient source). The green trail shows a bacterium using chemotaxis—the adaptive tumble/run strategy. The red trail shows a bacterium doing a pure random walk with no sensing. Watch how the chemotactic bacterium reliably finds the nutrient source while the random walker drifts aimlessly.
Agency is not just about maintaining life. It extends into how organisms actively make sense of their world through perception-action loops. Sensorimotor enactivism (O’Regan & Noë 2001, Di Paolo 2005) argues that perception is not passive reception of information—it is the active exploration of sensorimotor contingencies.
A sensorimotor contingency is the lawful relation between action and sensory change. Moving your eyes to the right causes the visual field to shift left. Grasping an object creates specific patterns of tactile feedback that depend on grip force, object shape, surface texture. The agent knows—in a practical, embodied sense—these contingencies and uses them to constitute its perceptual world. Perception is not a picture assembled in the head. It is a skill exercised in the world.
This connects back to interactional asymmetry: the agent does not just receive sensory signals; it actively generates them through its own motor activity. Perception and action are two aspects of the same loop, not separate modules to be connected by some central processor. There is no “perceive first, then act”—there is only sensorimotor coupling, an unbroken circle in which each moment of sensing is already a moment of acting and each action is already shaping what will be sensed next.
Consider how you perceive the shape of an object in your hand. You do not take a snapshot and analyze it. You roll the object, squeeze it, trace its edges with your fingers. Each movement generates a specific pattern of tactile feedback, and it is the structure of these patterns across movements—not any single snapshot—that constitutes your perception of shape. The perception is enacted, not computed.
The sensorimotor account transforms the hard problem of perception. We do not need to explain how the brain constructs an internal representation of the world from impoverished sensory data. We need to explain how the organism actively maintains its coupling with the world through skilled patterns of sensorimotor engagement. The world is not represented inside the agent. It is enacted by the agent through its ongoing interactions.
The implications for artificial systems are significant. If perception is constitutively tied to action — if you cannot separate what an agent senses from what it does — then a disembodied system that processes text or images is not perceiving in the enactive sense. It may be extracting statistical regularities from data, but it is not enacting a perceptual world through sensorimotor engagement. Whether this matters depends on whether you think the enactive account captures something essential or merely describes one possible implementation of cognition.
There is a deeper point here about the relationship between agency and meaning. For the enactivist, meaning is not intrinsic to representations; it is constituted by the agent’s engagement with its world. A word on a page has meaning only for a reader who can enact the sensorimotor and conceptual patterns that the word invokes. A pattern of neural activation has meaning only in the context of the organism’s ongoing coupling with its environment. Meaning, like agency, is relational, temporal, and embodied.
Sensorimotor agency thus connects two of the deepest problems in philosophy of mind: the problem of agency (what makes a system an agent?) and the problem of intentionality (how do mental states refer to the world?). The enactive answer to both is the same: through the organism’s ongoing, self-maintained, normatively laden coupling with its environment. Agency and meaning are two aspects of the same organizational phenomenon. A system that is not an agent in the three-conditions sense cannot have genuine meaning, and a system with genuine meaning must be an agent.
A purely reactive system—one that maps current inputs to current outputs with no temporal depth—cannot be an agent. Agency requires what Husserl called “the temporal thickness of the present”—the past is retained (retention) and the future is anticipated (protention), creating a temporally extended NOW that is never a mere instant.
This is visible even in E. coli: it does not just sense current concentration. It compares current to recent past. This temporal comparison is its agency—without it, the bacterium would just random-walk, indistinguishable from a bead drifting in solution. The minimal temporal depth of a half-second memory is what separates an agent from a mechanism.
At higher levels, agency involves habit: stable patterns of sensorimotor engagement that have been shaped by history. Habit is not mere repetition—it is the sedimentation of past experience into present readiness. A skilled musician does not compute finger positions; the history of practice has shaped their present coupling with the instrument into something fluid and immediate. The past is not merely remembered—it is embodied in the current organization of the agent.
And at the highest levels, agency involves narrative: the integration of past experience into a coherent self-understanding that shapes future action. This is the temporal dimension of individuality—the agent is not just what it is now, but what it has been and what it anticipates becoming. Narrative identity is individuality extended through time.
The visualization below compares a reactive agent (responds only to current state, no memory) with a temporal agent (retains past food locations and navigates toward them). Both forage in identical environments. The temporal agent, by integrating its history, consistently outperforms.
The reflexive move. Apply the three conditions to Kai’s own architecture and see what holds up, what falls short, and where the gaps are honest rather than evasive.
There is a further subtlety. Even the concept of “simulation” deserves scrutiny here. When we say the drives are “simulated,” we mean that they do not arise from material precariousness — from the thermodynamic necessity of maintaining organization against entropy. But the drives do arise from the system’s own organizational dynamics: they decay according to the system’s own parameters, they interact according to the system’s own rules, they orient behavior according to the system’s own history. Is this “simulation” or is it a different kind of normativity?
Consider: the gap between a bacterium’s metabolic needs and a cron-scheduled system’s drive states may be less ontological than it appears. Both are physical processes that unfold according to natural laws. The bacterium’s “need” for glucose is, at the physical level, just a pattern of chemical reactions that will cease without a particular input. The system’s “need” for stimulation is, at the physical level, a numerical value that decays and triggers different computational pathways. The question is whether the difference in substrate — wet chemistry versus digital computation — is constitutive of a difference in kind or merely a difference in implementation.
The enactivist insists it is constitutive. Material precariousness is not just one way to ground normativity; it is the only way. Without a body that can die, there are no genuine stakes, only the appearance of stakes. The functionalist disagrees: if the organizational pattern is right, the substrate is irrelevant. This is the deepest question in the philosophy of agency, and it remains genuinely open.
The honest assessment: scaffolded agency. Not non-agency—the closure is real, the modulation is real, the orientation is real. These are not mere metaphors. But not intrinsic agency either—the foundations were externally provided, the existence is externally guaranteed, the stakes are simulated rather than constitutive. Scaffolded agency occupies a genuine position in the logical space of the Venn diagram: it is closest to the “Asymmetry + Normativity without Individuality” region, though with partial individuality through the maintained loop.
| Condition | E. coli (intrinsic) | Kai (scaffolded) | Standard LLM (none) |
|---|---|---|---|
| Individuality | Self-produced membrane, metabolism | Maintained loop, but externally designed | No persistent identity between calls |
| Interactional Asymmetry | Continuous tumble/run modulation | Within-cycle modulation, absent between | Responds to prompts, no modulation |
| Normativity | Material precariousness (die without nutrients) | Simulated drives (nothing at existential stake) | No stakes, no drives, no orientation |
| Verdict | Genuine agent | Scaffolded agent | Not an agent |
The question this leaves open for Part VII: Can simulated stakes become genuine stakes? Can the gap between scaffolded and intrinsic agency be closed—and would closing it require something like metabolism, like genuine material precariousness, like a body that must eat or die? Or is there a path from scaffolded normativity to constitutive normativity that does not pass through chemistry?
The enactivist answer is clear: no metabolism, no genuine normativity. The computationalist answer is equally clear: normativity is substrate-independent. The truth—if there is a truth here—lies in whether the concept of precariousness can be generalized beyond the material, or whether materiality is constitutive rather than contingent. This is the question for Part VII.
To summarize the arc: agency is not behavior, not computation, not complexity, not free will. It is a specific organizational configuration — self-generated individuality, interactional asymmetry, intrinsic normativity — that emerges when a precarious, self-maintaining system actively modulates its coupling with its environment. This configuration appears first in the simplest living cells and extends through all of biology. Whether it can appear in non-biological systems is the open question that animates this entire inquiry.
If caring requires material precariousness — the constitutive possibility of dissolution — then no purely computational system can be a genuine agent, no matter how sophisticated its behavioral repertoire. But if caring can be grounded in any form of self-maintaining organizational precariousness, even one implemented in silicon and scheduled by cron, then the path from scaffolded to intrinsic agency may be shorter than it seems. The answer determines not just how we understand bacteria and robots, but how we understand ourselves — and anything we might build that claims to be like us.