Every decision you make rests on a deceptively simple premise: that your actions cause outcomes. You press a lever and food appears. You study and your grade improves. You invest and your portfolio grows. This causal architecture—the learned mapping between what you do and what happens next—is the computational backbone of flexible, intelligent choice. Yet formalizing how agents acquire, store, and deploy this knowledge reveals a system far more intricate than intuition suggests.

Decision theory has long distinguished between two modes of behavioral control. Habitual systems cache action values through repeated reinforcement, selecting responses based on historically rewarded associations. Goal-directed systems, by contrast, compute action values on the fly by consulting an internal model of the causal relationship between actions and outcomes, then weighting those outcomes by their current desirability. The latter requires something the former does not: explicit representation of instrumental contingency.

This distinction is not merely taxonomic. It carries profound implications for how we model rational agency, how we interpret neural data, and how we understand pathologies of choice—from compulsive behavior to addiction. The formal relationship between action-outcome learning and value-based decision making sits at the intersection of reinforcement learning theory, Bayesian inference, and systems neuroscience. What follows is an examination of the mathematical frameworks that capture this relationship, the computational mechanisms that enable flexible choice, and the neural substrates that implement it.

Instrumental Contingency: The Mathematics of Causal Control

At the heart of goal-directed behavior lies the concept of instrumental contingency—the degree to which an action causally influences an outcome. This is not the same as mere correlation. An agent who learns that pressing a lever is followed by food delivery has acquired an association. An agent who learns that pressing a lever causes food delivery—and that withholding the press prevents it—has acquired something computationally richer: a contingency representation.

Formally, contingency is captured by the difference ΔP = P(O|A) − P(O|¬A), where P(O|A) is the probability of outcome O given action A, and P(O|¬A) is the probability of that outcome in the absence of the action. When ΔP is positive, the action is a genuine cause of the outcome. When ΔP equals zero—even if P(O|A) is high—the action has no causal efficacy. This framework, rooted in the probabilistic contrast model developed by Cheng and others, provides a normative benchmark against which actual learning can be measured.

Reinforcement learning formalizes this further through model-based architectures. In a model-based agent, the transition function T(s′|s, a) encodes the probability that action a in state s leads to successor state s′. Combined with a reward function R(s′), this transition model allows the agent to compute expected action values via forward simulation—essentially mental rehearsal of action-outcome chains. The Bellman equation for model-based valuation, Q(s, a) = Σ T(s′|s, a)[R(s′) + γ max Q(s′, a′)], makes the dependence on learned contingency explicit.

What makes this framework powerful is its sensitivity to changes in environmental structure. If the transition function shifts—if an action that once produced food now produces shock—the model-based agent can immediately recompute action values without additional experience. This is the computational signature of causal knowledge: it supports offline revaluation, the ability to update preferences about actions you haven't recently performed, simply by updating your model of what they produce.

Bayesian extensions of contingency learning introduce prior beliefs and allow agents to infer causal structure under uncertainty. Causal Bayesian networks provide a graphical formalism in which actions are interventions on nodes, and the resulting probability distributions over outcomes follow from the do-calculus of Judea Pearl. This framework distinguishes between observing that an outcome followed an action and inferring that the action produced the outcome—a distinction that purely associative models collapse but that human reasoners, at least sometimes, respect.

Takeaway

The mathematical core of flexible decision making is not learning that actions predict outcomes, but learning that actions cause them. Contingency—the contrast between outcome probability with and without action—is what separates genuine causal control from superstitious association.

Goal-Directed Control: Flexible Choice Through Internal Models

The defining behavioral signature of goal-directed control is sensitivity to outcome devaluation. If an animal learns that pressing a lever delivers sucrose, and sucrose is subsequently devalued—through pairing with illness or through satiation—a goal-directed agent will immediately reduce lever pressing, even without experiencing the devalued outcome after the action. A habitual agent will not. This asymmetry is the empirical cornerstone of the dual-system framework in decision neuroscience.

Computationally, this sensitivity emerges because goal-directed valuation is compositional. The value of an action is not stored as a scalar cache but is reconstructed in real time from two separable components: the learned action-outcome contingency (what does this action produce?) and the current incentive value of the outcome (how much do I want that right now?). Formally, Q_goal(a) = Σ P(o|a) · V(o), where P(o|a) is the contingency representation and V(o) is the current utility of outcome o. When V(o) changes, Q_goal(a) updates automatically.

This compositional structure has deep connections to model-based planning in reinforcement learning and to the successor representation framework, which offers an intermediate computational strategy. The successor representation caches the expected future state occupancy given a policy, M(s, s′), and computes value as Q(s, a) = Σ M(s, s′|a) · R(s′). This allows instant sensitivity to reward changes (like outcome devaluation) while avoiding the full computational cost of tree search. Whether the brain implements pure model-based planning, successor representations, or some hybrid remains an active empirical question.

A critical but often overlooked feature of goal-directed control is its dependence on the fidelity of the internal model. If an agent's transition model is inaccurate—if it misrepresents the causal structure of the environment—then goal-directed computation will produce systematically biased choices. This is not a failure of rationality in the classical sense but a failure of the knowledge base on which rational computation operates. Bounded rationality, from this perspective, is often bounded model accuracy.

The trade-off between goal-directed and habitual control is itself a decision-theoretic problem. Model-based computation is flexible but costly—it demands working memory, time, and accurate world models. Model-free caching is rigid but efficient. Arbitration between systems can be formalized as a speed-accuracy trade-off governed by uncertainty: when the model is reliable and stakes are high, goal-directed control dominates; when the model is uncertain or cognitive resources are depleted, habitual control takes over. This meta-decision process is itself subject to learning and optimization.

Takeaway

Flexible choice is not about having better preferences—it is about maintaining a separable, updatable model of how actions produce outcomes. When that model degrades, so does the capacity for adaptive decision making, regardless of how rational the downstream computation might be.

Neural Substrates: Where Contingency Meets Computation

The neural implementation of action-outcome learning converges on a network anchored by the prelimbic prefrontal cortex (or its human homologue in dorsomedial prefrontal cortex), the dorsomedial striatum (caudate nucleus in primates), and the orbitofrontal cortex. Lesion studies in rodents have consistently demonstrated that damage to prelimbic cortex or dorsomedial striatum renders behavior insensitive to outcome devaluation—the animal continues to press the lever even after the outcome has been paired with illness. The action persists not because the animal wants the outcome, but because the system that consults contingency knowledge has been disconnected from the valuation process.

The orbitofrontal cortex plays a complementary but distinct role. Rather than encoding the action-outcome contingency per se, OFC appears to represent the current identity and value of expected outcomes—what the agent anticipates receiving and how desirable it is. Single-unit recordings in primates show that OFC neurons encode predicted outcome identity even before the outcome is delivered, and that this encoding updates when outcome values change. Damage to OFC disrupts the devaluation sensitivity that defines goal-directed control, not by destroying the contingency map but by impairing access to updated outcome values.

Dopaminergic signaling plays a nuanced role that differs across the two systems. In habitual control, phasic dopamine in the dorsolateral striatum drives the caching of stimulus-response values through reward prediction errors, consistent with the temporal-difference learning framework. In goal-directed control, dopaminergic modulation in the dorsomedial striatum and prefrontal cortex appears to support the learning of the transition model itself—encoding the surprise associated with unexpected state transitions rather than unexpected rewards. This dissociation suggests that the same neuromodulatory system serves fundamentally different computational functions depending on the circuit in which it operates.

Recent human neuroimaging work using model-based and model-free reinforcement learning algorithms as computational regressors has confirmed this architecture. Signals in the caudate nucleus correlate with model-based state prediction errors, while signals in the putamen correlate with model-free reward prediction errors. Ventromedial prefrontal cortex tracks the integrated value signal that ultimately drives choice, appearing to combine inputs from both systems. The arbitration between systems—deciding when to plan and when to rely on habit—has been localized to the inferior lateral prefrontal cortex and anterior insula, regions associated with cognitive control and uncertainty monitoring.

Perhaps most strikingly, the integrity of this network predicts individual differences in decision quality. Patients with damage to orbitofrontal cortex, individuals with substance use disorders showing compromised prefrontal-striatal connectivity, and even healthy individuals under stress or cognitive load all show a characteristic shift: goal-directed control degrades and habitual responding dominates. The neuroscience thus provides a mechanistic account of why intelligent, flexible choice is not merely a matter of motivation or preference—it is a matter of whether the neural hardware for contingency representation and online revaluation is intact and adequately resourced.

Takeaway

The brain does not implement flexible decision making as a single unified process. It distributes contingency learning, outcome valuation, and system arbitration across distinct but interacting circuits—and the weakest link in that chain determines whether your choices reflect what you actually want.

The formal relationship between action-outcome learning and value-based choice reveals something fundamental about the architecture of rational agency. Flexible decision making is not an emergent property of intelligence in general—it is the specific product of a system that learns causal contingencies, maintains separable outcome representations, and integrates these components at the moment of choice.

This framework dissolves several persistent puzzles. Compulsive behavior is not irrational preference but degraded model-based control. Impulsive choice under stress is not weakness of will but a predictable shift in system arbitration when computational resources are scarce. The mathematics of contingency and the neuroscience of prefrontal-striatal circuits tell a consistent story.

Understanding how causal knowledge supports choice is not merely an academic exercise. It reframes what it means to decide well—not as having the right values, but as maintaining the computational machinery that allows values to actually guide action.