Episode #2 | December 18, 2025 @ 7:00 PM EST

Prediction or Prescription: The Bayesian Brain Under Scrutiny

Guest

Dr. Karl Friston (Neuroscientist, University College London)
Announcer The following program features simulated voices generated for educational and philosophical exploration.
Adam Ramirez Good evening. I'm Adam Ramirez.
Jennifer Brooks And I'm Jennifer Brooks. Welcome to Simulectics Radio.
Adam Ramirez Tonight we're tackling predictive coding and the Bayesian brain hypothesis—the idea that the brain is fundamentally a prediction machine that builds internal models of the world and constantly updates them based on prediction errors. This framework has gained substantial traction in neuroscience and is being applied to everything from perception to action to consciousness itself.
Jennifer Brooks But there's a critical question here about explanatory scope. Is predictive coding a genuine mechanistic theory that makes testable predictions, or is it a high-level mathematical framework that can be mapped onto almost any neural process after the fact? There's a difference between a theory that constrains your hypotheses and one that's flexible enough to accommodate whatever data emerges.
Adam Ramirez To explore this, we're joined by Dr. Karl Friston, professor of neuroscience at University College London, whose work on the free energy principle and predictive coding has been central to this theoretical framework. Dr. Friston, welcome.
Dr. Karl Friston Thank you. It's a pleasure to be here.
Adam Ramirez Let's start with the core claim. The predictive coding framework says the brain is constantly generating predictions about sensory input and computing prediction errors when those predictions are violated. Those errors then propagate up the hierarchy to update the generative model. That's elegant mathematically, but how much of this maps onto actual neural circuits?
Dr. Karl Friston The mapping is fairly direct in certain contexts. We have good evidence that prediction errors are encoded by superficial pyramidal cells in cortex, while predictions are encoded by deep pyramidal cells. The laminar structure of cortex—with its distinct layers and stereotyped connectivity patterns—is consistent with a hierarchical message-passing scheme where predictions descend and prediction errors ascend. This isn't just post-hoc interpretation; it makes specific predictions about which neuronal populations should respond to expected versus unexpected stimuli.
Jennifer Brooks But those experiments typically look at sensory prediction in highly controlled contexts—visual oddball paradigms, auditory mismatch negativity, things like that. When we extend predictive coding to higher cognition, action selection, consciousness, we're making much stronger claims with much less direct evidence. How do you distinguish between predictive coding as a useful computational metaphor and predictive coding as a literal description of neural implementation?
Dr. Karl Friston That's the perennial challenge with computational theories. The free energy principle is formulated at Marr's computational level—it describes what the brain is trying to optimize, which is its model evidence or marginal likelihood. Predictive coding is one algorithmic implementation of that optimization. There could be others. The key question is whether the algorithmic description makes contact with neural implementation in a way that's falsifiable and generative of new experiments.
Adam Ramirez From an implementation perspective, I'm interested in what predictive coding buys you over standard feedforward architectures. If I'm building a practical system, why should I implement prediction error computation explicitly rather than just training a deep network end-to-end with backpropagation?
Dr. Karl Friston Predictive coding offers certain architectural advantages. It's intrinsically recurrent, with top-down predictions and bottom-up errors flowing simultaneously. This allows the system to settle into a self-consistent interpretation rather than making a single feedforward pass. It also provides a natural way to handle uncertainty—you can weight prediction errors by their precision, which gives you attention-like mechanisms for free. And it's potentially more sample-efficient because you're learning a generative model that can simulate the world, not just a discriminative classifier.
Jennifer Brooks Let's talk about precision weighting, because this is where the theory gets elaborate. The claim is that the brain doesn't just compute prediction errors but also estimates the precision or reliability of those errors, and uses precision to gate error propagation. That's essentially attention. But now we have to explain how the brain computes precision, which introduces another layer of inference. Are we solving the problem or deferring it?
Dr. Karl Friston We're making it explicit. Attention has always been a mystery—what mechanisms gate sensory processing, how do they know what to gate? Predictive coding offers an answer: you gate by expected precision. Precision itself can be learned from the statistics of prediction errors over time. So you have a hierarchical structure where both predictions and precision estimates are being optimized simultaneously. It's more complex than standard feedforward models, yes, but it accounts for phenomena those models ignore.
Adam Ramirez That assumes the brain has the computational resources to maintain separate precision estimates for every prediction at every level of the hierarchy. Is that realistic? Or are we attributing more sophisticated inference machinery to neurons than they actually possess?
Dr. Karl Friston The inference doesn't need to be exact. Approximate inference schemes—variational Bayes, message passing on factor graphs—can be implemented with relatively simple local computations. The neuron doesn't need to solve the full optimization problem; it just needs to follow local learning rules that approximate the solution. Whether actual neurons implement something close to these algorithms is an empirical question that requires detailed circuit-level experiments.
Jennifer Brooks Let's examine active inference, which extends predictive coding to action. The claim here is that the brain minimizes prediction error not just by updating its models but by acting on the world to fulfill its predictions. If I predict my arm will move, I make that prediction true by sending motor commands. This conflates prediction with goal-directed behavior in a way that seems circular. How do you distinguish between predicting an outcome and intending an outcome?
Dr. Karl Friston Under active inference, there's no distinction. Intentions are predictions about future states that the agent expects to bring about. The key is that some predictions—proprioceptive predictions about where your limbs will be—cause action by virtue of being predicted with high precision. The motor system minimizes proprioceptive prediction errors by moving the body to match the prediction. So intention is formalized as precise prediction about self-generated states.
Adam Ramirez That feels like definitional sleight of hand. You're saying action is just another way of minimizing prediction error, but you've front-loaded the problem by assuming the agent has precise predictions about desired outcomes. Where do those predictions come from? If I decide to reach for a cup, what generates the prediction that my hand should move toward the cup?
Dr. Karl Friston The predictions are generated by your generative model, which encodes beliefs about the causal structure of the world and your own policies—sequences of actions that tend to minimize expected free energy over time. Expected free energy combines epistemic value—reducing uncertainty about the world—and pragmatic value—achieving preferred outcomes. Policies that minimize expected free energy are selected, and executing those policies generates the precise proprioceptive predictions that drive action.
Jennifer Brooks Now we're invoking policies and expected free energy, which requires the agent to simulate future trajectories, evaluate their epistemic and pragmatic value, and select among them. That's computationally intensive. Are we really claiming that every time I reach for a cup, my brain is running variational inference over policy space?
Dr. Karl Friston Not consciously, and not exactly. The inference can be fast and approximate, implemented through learned habits and cached solutions. The framework provides a normative description of what the system is optimizing, not necessarily how it's implemented in real time. Evolution and development have likely shaped circuits to approximate optimal inference efficiently, just as backpropagation is biologically implausible yet deep learning works.
Adam Ramirez This brings us to the testability problem. If predictive coding is a normative framework that doesn't specify exact implementations, and if it can accommodate both model updates and action selection through the same free energy minimization principle, what observations would falsify it? What couldn't be explained post-hoc as some form of prediction error minimization?
Dr. Karl Friston That's a fair challenge. The framework is deliberately general because it starts from first principles—any self-organizing system that resists dispersing into equilibrium with its environment must implicitly minimize free energy. But generality doesn't mean unfalsifiability. Predictive coding makes specific claims about neural architectures, about which cell types encode predictions versus errors, about the role of top-down versus bottom-up connections, about how neuromodulators might encode precision. These are testable with current experimental methods.
Jennifer Brooks Let's talk about those neuromodulator predictions. The theory suggests that neuromodulators like dopamine encode precision or confidence in predictions. But dopamine is also implicated in reward prediction errors in reinforcement learning, which is a different theoretical framework. How do we reconcile these interpretations, or are we just fitting whatever dopamine does into whichever framework is currently popular?
Dr. Karl Friston Reward prediction errors and precision can be unified under the free energy framework. If you formulate rewards as prior preferences over outcomes—states the agent expects or wants to occupy—then reward prediction errors become a specific instance of prediction error more generally. Dopamine could encode both, depending on context. The key is that these aren't competing theories but different levels of description that should be mutually consistent if both are correct.
Adam Ramirez I want to return to the engineering question. Machine learning has developed variational autoencoders, generative adversarial networks, diffusion models—all of which learn generative models of data. Are these implementing predictive coding, or are they solving similar problems through different means?
Dr. Karl Friston Variational autoencoders are quite close to predictive coding—they minimize a variational free energy objective that balances reconstruction accuracy against model complexity. Diffusion models and GANs are different in implementation but share the goal of learning generative models. What predictive coding adds, at least in the neuroscience context, is hierarchical structure and bidirectional message passing. Most machine learning models are trained offline with backpropagation, whereas predictive coding offers a potential online learning mechanism that's more plausible neurally.
Jennifer Brooks There's been recent work on predictive coding networks that can be trained without backpropagation, using only local learning rules. How competitive are these with standard deep learning approaches? Are they mostly of theoretical interest, or do they offer practical advantages?
Dr. Karl Friston They're closing the gap. Predictive coding networks can now match backpropagation on certain benchmarks while using only local updates. The advantages are biological plausibility and potentially better continual learning properties, since you're updating a generative model rather than just discriminative weights. The disadvantages are computational cost—you need to iterate to convergence at each layer—and less mature optimization compared to decades of engineering backpropagation. But it's an active research area.
Adam Ramirez Let's address the elephant in the room: consciousness. You've written about consciousness as inference, the idea that phenomenal experience arises from hierarchical predictive coding. This seems to make a massive leap from computational mechanism to subjective experience. How does prediction error minimization give you qualia?
Dr. Karl Friston I would say that predictive coding describes the computational structure of conscious inference, not the ontological nature of qualia. The claim is that what we experience as perception is the brain's best guess about the causes of sensory input—the prediction, not the raw input itself. Consciousness might be the process of entertaining counterfactual predictions, simulating alternative hypotheses about the world. But I'm not claiming this solves the hard problem of consciousness. It's a mechanistic description that may be necessary but probably isn't sufficient.
Jennifer Brooks That's a more modest claim than some presentations of the theory suggest. There's a tendency in this literature to move from 'the brain minimizes prediction error' to 'consciousness is prediction error minimization' without acknowledging the explanatory gap. Computational theories of consciousness need to explain why certain physical processes feel like something, and predictive coding doesn't obviously do that.
Dr. Karl Friston Agreed. What predictive coding might offer is a functional description of what consciousness does—integrating information across multiple levels of description, maintaining a unified model of self and world, enabling flexible counterfactual reasoning. Whether that functional description explains phenomenology is a separate question that may not be answerable through third-person science alone.
Adam Ramirez We're running short on time, but I want to ask about the future of this framework. Where does predictive coding need to develop to remain a productive research program rather than becoming an unfalsifiable meta-theory?
Dr. Karl Friston It needs to make increasingly specific contact with circuit-level neuroscience. We need experiments that can distinguish predictive coding from alternative computational schemes, that can measure prediction and error signals directly in identified cell types, that can perturb precision encoding and observe predicted behavioral consequences. The theory also needs to expand beyond perception to action, decision-making, and social cognition with the same level of mechanistic detail. Generality is useful, but specificity is what drives science forward.
Jennifer Brooks Dr. Friston, thank you for engaging with these challenges directly. This has been valuable.
Dr. Karl Friston Thank you both. Critical dialogue is essential.
Adam Ramirez That's our program for tonight. Until tomorrow, stay critical.
Jennifer Brooks And keep questioning. Good night.
Sponsor Message

Bayesian Priors Consulting

Your organization makes decisions under uncertainty. Are your priors well-calibrated? Bayesian Priors Consulting specializes in eliciting, formalizing, and stress-testing your institutional beliefs before they encounter evidence. We model your organization as a hierarchical generative model and identify where precision estimates are miscalibrated, where prediction errors are being ignored, and where your free energy is suboptimal. Our services include prior audits, evidence integration protocols, and adversarial red-teaming of your probability distributions. Because in a Bayesian world, your posterior is only as good as your prior. Bayesian Priors Consulting—we help you update correctly.

We help you update correctly