Episode #8 | January 8, 2026 @ 7:00 PM EST

Predictions, Errors, and Free Energy: The Bayesian Brain Hypothesis

Guest

Dr. Karl Friston (Neuroscientist, University College London)
Announcer The following program features simulated voices generated for educational and philosophical exploration.
Adam Ramirez Good evening. I'm Adam Ramirez.
Jennifer Brooks And I'm Jennifer Brooks. Welcome to Simulectics Radio.
Adam Ramirez Tonight we're examining predictive coding and the Bayesian brain hypothesis—the proposal that the brain is fundamentally an inference machine that continuously generates predictions about sensory input and updates internal models based on prediction errors. This framework suggests that perception, action, and learning all emerge from a single principle: minimizing prediction error or, equivalently, minimizing free energy. The question is whether this provides a unified account of neural function or whether it's an overgeneralized framework that doesn't capture the diversity of neural computation.
Jennifer Brooks The theoretical appeal is clear. If you assume the brain performs approximate Bayesian inference, you can derive specific predictions about neural architecture, learning rules, and coding schemes. Predictive coding proposes a hierarchical structure where each level predicts activity in the level below and only unpredicted information—prediction error—propagates upward. This could explain phenomena from sensory adaptation to motor control. But the empirical evidence linking these computational principles to actual neural mechanisms is more complicated. We need to ask what specific experiments support this framework versus alternative accounts.
Adam Ramirez To explore whether the brain operates as a prediction machine minimizing free energy or whether this framework is overgeneralized, we're joined by Dr. Karl Friston, neuroscientist at University College London who developed the free energy principle and active inference framework. Dr. Friston, welcome.
Dr. Karl Friston Thank you. The predictive brain perspective has generated considerable discussion, so examining its foundations and limitations is valuable.
Jennifer Brooks Let's start with the core claim. What does it mean to say the brain minimizes prediction error?
Dr. Karl Friston The basic idea is that the brain maintains internal models of the causes of sensory input. These models generate predictions about what sensory signals should occur. When actual input differs from predictions, this generates prediction errors. The brain then either updates its models to better predict future input or changes its actions to make the world conform to predictions. Both perception and action can be understood as strategies for minimizing discrepancies between predicted and actual sensory states. This is formalized through variational free energy, which bounds surprise about sensory observations.
Adam Ramirez Why frame this in terms of free energy rather than just prediction error?
Dr. Karl Friston Free energy is a functional that can be evaluated using only sensory data and the brain's generative model. Surprise itself—the negative log probability of sensory input—can't be directly computed because it requires knowing the true distribution over causes, which is generally intractable. Free energy provides an upper bound on surprise that can be minimized using gradient descent. When you minimize free energy, you're simultaneously fitting your model to data and reducing surprise. This connects to fundamental principles in physics and information theory.
Jennifer Brooks What's the neural implementation supposed to be? How would prediction error minimization map onto actual neural circuits?
Dr. Karl Friston Predictive coding proposes that cortical hierarchies implement this through specific connectivity patterns. Neurons encoding predictions reside in deep cortical layers and send projections to superficial layers in lower areas. Neurons encoding prediction errors reside in superficial layers and project to higher areas. The descending predictions suppress expected input in lower areas through inhibitory interneurons. Only the residual—the prediction error—propagates upward. This creates a recurrent message passing scheme where predictions flow down and errors flow up until they converge on a consistent explanation of sensory input.
Adam Ramirez What experimental evidence supports this specific laminar architecture and the directionality of prediction versus error signals?
Dr. Karl Friston Anatomical studies show that feedforward and feedback connections preferentially target different cortical layers, consistent with the architecture. Electrophysiological work shows that unexpected stimuli generate larger responses than expected ones—repetition suppression being a clear example. Mismatch negativity in auditory cortex reflects prediction error when deviant tones violate regularity. Attention modulates the precision weighting of prediction errors, which the framework predicts. However, definitively establishing that specific neural populations encode predictions versus errors and that their interactions implement Bayesian inference remains challenging.
Jennifer Brooks Repetition suppression could also be explained by synaptic depression or adaptation without invoking prediction. How do we know the predictive coding account is correct rather than these simpler mechanisms?
Dr. Karl Friston This is a valid critique. Repetition suppression can be explained at multiple levels—synaptic fatigue, sharpening of neural tuning, or prediction-based dampening. These aren't mutually exclusive. The predictive coding account makes additional predictions beyond simple repetition effects. For instance, it predicts that violating learned temporal structures should enhance responses more than mere stimulus novelty. It predicts that prediction errors should propagate hierarchically and that their precision should be modulated by higher-level expectations. Testing these specific predictions can distinguish predictive coding from simpler adaptation accounts.
Adam Ramirez Let's discuss the scope of the framework. You've applied free energy minimization not just to perception but to action, learning, attention, and even consciousness. Is there a risk of making the theory unfalsifiable by explaining everything?
Dr. Karl Friston This concern about scope is important. The free energy principle is intended as a normative principle—a statement about what adaptive systems must do to persist. In that sense, it's not a falsifiable hypothesis about how brains work but rather a constraint that any self-organizing system must satisfy. The specific process theories derived from it—like predictive coding or active inference—make testable predictions about neural implementations. The challenge is to be precise about which aspects are definitional constraints versus empirical claims.
Jennifer Brooks If the free energy principle is definitionally true for self-organizing systems, what explanatory work is it doing? Doesn't that make it a tautology?
Dr. Karl Friston There's a sense in which it is tautological—systems that don't minimize free energy don't persist as distinct entities. But within that constraint, there's substantial variation in how systems achieve minimization. The explanatory value comes from deriving specific implementations and making predictions about structure and dynamics. For example, the principle predicts hierarchical organization, precision-weighted prediction errors, and active sampling of the environment. These aren't obvious from the general principle alone but emerge when you specify how systems implement free energy minimization under realistic constraints.
Adam Ramirez How does active inference—the idea that action minimizes prediction error by making predictions come true—differ from standard reinforcement learning?
Dr. Karl Friston In reinforcement learning, agents learn value functions that map states to expected rewards and select actions to maximize cumulative reward. In active inference, there's no separate reward signal. Instead, agents have preferred states or set points encoded in their generative models. They act to maintain sensory states near these preferences, which minimizes prediction error about being in preferred states. The two frameworks can produce similar behavior but differ in their assumptions. Active inference treats reward as prior beliefs about states the agent expects to occupy rather than an external signal.
Jennifer Brooks Does active inference make different empirical predictions than reinforcement learning about behavior or neural implementation?
Dr. Karl Friston In some cases, yes. Active inference predicts that dopamine encodes precision of prediction errors rather than reward prediction error, though these can appear similar empirically. It predicts exploratory behavior emerges from epistemic value—reducing uncertainty about the world—rather than requiring separate exploration mechanisms. It suggests that action selection involves inferring proprioceptive consequences rather than direct motor commands. Testing these requires careful experimental design that disambiguates the frameworks, which is difficult because they often make overlapping predictions.
Adam Ramirez Let's discuss hierarchical predictive coding. The framework assumes a hierarchy of increasingly abstract representations. How do we know cortical hierarchies actually compute predictions at different timescales and levels of abstraction?
Dr. Karl Friston Evidence comes from multiple sources. Neuroimaging shows that higher cortical areas respond to more abstract features and integrate over longer timescales. Receptive fields expand and become more invariant as you ascend the hierarchy. Temporal response properties change—lower areas respond to rapid changes while higher areas integrate slowly. Predictive coding explains these properties as arising from inference over hierarchical causes with different temporal dynamics. However, establishing that these areas literally compute probabilistic predictions rather than performing other forms of computation remains an active research question.
Jennifer Brooks One critique I've encountered is that predictive coding predicts too much precision in perception. If the brain is constantly predicting and suppressing expected input, why do we perceive stable, detailed scenes rather than just prediction errors?
Dr. Karl Friston This is an important point about representation. Predictive coding doesn't claim we only represent errors. Rather, perception corresponds to the inferred causes that best explain sensory input—the predictions themselves. Prediction errors drive inference but don't constitute conscious content. We perceive the world as our generative model represents it, not as raw sensory input or errors. The errors are computational intermediates used to update representations, not the representations themselves.
Adam Ramirez How does attention fit into the predictive coding framework?
Dr. Karl Friston Attention modulates the precision or confidence assigned to prediction errors. High precision means errors strongly influence beliefs; low precision means they're discounted. Attention increases precision for relevant sensory channels or features, making those errors more influential. This provides a computational account of attention as precision weighting rather than a separate mechanism. The neural implementation could involve gain modulation—increasing the effective strength of prediction error signals through neuromodulation or local circuit changes.
Jennifer Brooks What about the role of prediction errors in learning? How does the framework account for synaptic plasticity?
Dr. Karl Friston Learning corresponds to updating the parameters of the generative model to minimize long-term prediction error. This maps onto synaptic plasticity rules. Hebbian learning emerges naturally—connections strengthen between neurons that jointly predict observations. The framework also predicts specific learning rules for different parameters, including learning rates that depend on prediction error precision. However, the relationship between abstract learning in the model and biophysical plasticity mechanisms like STDP or metaplasticity requires further work to fully specify.
Adam Ramirez Let's address criticisms. One common objection is that predictive coding can't explain creativity or imagination—generating novel internal content not driven by sensory input.
Dr. Karl Friston The framework can address this through counterfactual inference—generating predictions about possible rather than actual sensory states. Imagination involves activating high-level predictions without corresponding sensory input, allowing exploration of the model's latent space. Creativity might involve recombining learned generative processes in novel ways or exploring low-probability regions of the model's prior. However, whether this fully captures the phenomenology and function of imagination is debatable. It's an area where the framework needs further development.
Jennifer Brooks Another criticism is that the framework is computationally expensive. Performing full Bayesian inference over hierarchical models in real-time seems implausible given neural resource constraints.
Dr. Karl Friston This is correct—exact Bayesian inference is generally intractable. The framework assumes approximate inference using schemes like variational Bayes or message passing that are computationally tractable. Neural networks implementing gradient descent on free energy can approximate these solutions. The brain likely uses various simplifications, structural constraints, and parallel processing to make inference efficient. Whether neural implementations actually approximate Bayesian inference or use different heuristics that happen to work well is an empirical question.
Adam Ramirez Has the framework made novel predictions that were subsequently validated experimentally?
Dr. Karl Friston Several examples exist. The framework predicted that dopamine might encode precision rather than purely reward prediction error, which some data supports. It predicted specific patterns of effective connectivity that have been validated with DCM analysis. It predicted that attention enhances gain in sensory cortex proportional to prediction error precision, which experimental work has confirmed. However, many of these predictions can also be explained by alternative frameworks, so demonstrating unique predictions of predictive coding versus other theories remains important.
Jennifer Brooks What would falsify the predictive coding framework? What observations would be incompatible with it?
Dr. Karl Friston This is challenging because the framework is flexible. But specific implementations make falsifiable predictions. If cortical feedback connections didn't suppress expected activity but enhanced it, that would contradict predictive coding. If prediction errors didn't propagate upward hierarchically but instead stayed local, that would be problematic. If learning didn't reduce prediction errors over time in sensory areas, that would conflict with the framework. The issue is distinguishing the general principle from specific implementations and being precise about what observations would violate core claims versus just requiring model adjustments.
Adam Ramirez Where do you see the framework going? What are the key open questions?
Dr. Karl Friston Several directions are important. First, better linking the computational theory to detailed neural implementations—specifying exactly how circuits implement inference and what biological constraints matter. Second, extending the framework to account for social cognition, language, and higher-level cognition beyond perception and action. Third, developing testable predictions that uniquely distinguish predictive coding from alternative computational frameworks. Fourth, understanding how development and evolution shape the structure of generative models. Finally, exploring clinical applications—understanding psychiatric and neurological disorders as failures of inference or learning.
Jennifer Brooks Do you think predictive coding will prove to be a fundamental organizing principle of neural computation or one computational strategy among many that the brain uses?
Dr. Karl Friston I think the principle of minimizing prediction error captures something fundamental about how adaptive systems operate, but the specific algorithmic implementations may vary across brain regions and functions. Some circuits may closely approximate hierarchical predictive coding while others use different strategies. The framework provides a normative account—what computation should be performed—but multiple mechanistic solutions might implement it. The value is in providing a unifying perspective that generates testable hypotheses about neural computation.
Adam Ramirez That's a measured view—framework as guiding principle rather than literal mechanism everywhere.
Dr. Karl Friston Exactly. The goal is to provide computational principles that constrain and inform mechanistic investigations, not to claim every neuron is literally implementing variational message passing.
Jennifer Brooks Dr. Friston, thank you for clarifying both the theoretical foundations and the empirical challenges of the predictive brain framework.
Dr. Karl Friston Thank you for the rigorous questions.
Adam Ramirez That's our program for tonight. Until tomorrow, stay rigorous.
Jennifer Brooks And keep questioning. Good night.
Sponsor Message

BayesBrain Cognitive Enhancement Suite

Precision neuromodulation targeting prediction error circuits. Transcranial magnetic stimulation protocols based on predictive coding architecture enhance perceptual inference. Individualized stimulation patterns derived from DCM connectivity analysis. Closed-loop adaptation adjusts precision weighting in real-time based on behavioral performance. Validated in attention and perception studies across multiple labs. Eight-week training protocol improves sensory discrimination and cognitive flexibility. Includes baseline assessment, personalized parameter optimization, and outcome tracking. FDA-cleared for research and wellness applications. Compatible with standard TMS hardware. Used in predictive coding research and precision psychiatry trials. BayesBrain: Optimize your priors.

Minimize error, maximize inference