We tend to picture thinking as content—beliefs, memories, sensations—rushing along a private wire. But there’s a second wire braided through the first: the mind noticing its own activity, estimating how noisy the line is, whether to continue, whether to switch channels. That second wire is metacognition. And the most honest language for it may be informational: entropy to index uncertainty, mutual information to track learning, compression to preserve structure with fewer bits. Not “dataism.” A quieter claim. Minds are finite channels living inside shifting environments; to persist, they must budget attention, select codes, and monitor the value of thinking-about-thinking as it happens.
Signals, Uncertainty, and the Cost of Knowing That You Know
Start with a basic picture: a signal travels through a noisy channel. Shannon’s insight wasn’t mystical; it was practical. There is a measurable ceiling on how much reliable information can pass. Minds live under the same ceiling. Perception, memory, language—each is a channel with capacity limits and characteristic noise. Now layer on a metacognitive monitor that estimates current uncertainty and chooses actions accordingly: seek more evidence, halt deliberation, switch tasks, or downgrade confidence. This second-order control is where information theory and metacognition fuse in practice: entropy estimates guide resource allocation; confidence is a code for expected error; attention reallocates bandwidth to reduce uncertainty where it most improves downstream loss.
Consider reading a technical paper. Your first pass is lossy; you skim headings and equations to sketch the distribution of ideas. Entropy stays high. A second pass collapses uncertainty around key claims. You pause at a proof, feel confusion spike (surprise, in Bayesian terms), then decide whether the value of more processing justifies the time. That decision itself is a signal. If your metacognitive entropy estimate is accurate, you avoid both overthinking (waste) and premature closure (error). If not, you get the familiar mess: rabbit holes, brittle conclusions, confident mistakes.
Two concrete tools matter here. First, mutual information—I(X;Y)—as a measure of how much “the text” reduces uncertainty about “the question I care about.” Metacognition isn’t maximizing information in general; it’s maximizing task-relevant information per unit cost. Second, redundancy. Skilled reasoners build deliberate redundancy in their internal codes: multiple cues (verbal, spatial, procedural) for the same concept; cross-checks between models; rephrasings out loud. Redundancy, paradoxically, improves throughput under noise. The metacognitive move is deciding where redundancy adds safety and where it bloats the code.
This is not abstract. In chess, strong players don’t only calculate lines; they sense when calculation stops paying off and switch to principles. In medicine, expert clinicians don’t merely accumulate tests; they know when a test’s expected information gain falls below its cost and potential harm. In daily planning, the best “productivity hack” is often a stop rule: no further research unless new information is likely to change the decision. That’s metacognitive capacity control, not willpower. If consciousness is a local reception point rather than a sealed object, then the quality of the reception—its precision weighting, its dynamic thresholding—becomes the locus of improvement. You don’t need a grand theory of mind to get traction. You need cleaner estimates of uncertainty, and affordable ways to lower it when it matters.
For a deeper dive from first principles—pattern, constraint, and mind as receiver—see information theory and metacognition.
Compression, Narrative, and the Temporary Self
Brains compress. They must. The world is broader than memory, so we keep summaries: habits, schemata, shortcuts. Good compression preserves structure—predictive regularities—while discarding idiosyncratic noise. The self, on this view, is not an origin point; it’s a rolling summary, a temporary compression of experience into a tractable model that keeps predicting well enough to avoid catastrophe. Metacognition then is the dynamic editor: which segments to re-encode at higher fidelity, which to downsample, when to rewrite the index entirely.
Information theory supplies a few handles. Minimum Description Length (MDL) balances model complexity against fit: a shorter code that still predicts wins. Minds obey a lived MDL. We tell stories that fit many observations with few moving parts. We resist updates that would explode the code length of our identities. That resistance is not mere stubbornness; it’s computational thrift. But thrift turns to trap if the world has changed and the code refuses to. The metacognitive task is to notice when the compression ratio has gone pathological—overly lossy, brittle under surprise—and choose to pay the extra bits now to regain flexibility later.
Time complicates this. Sequence feels like a river, but much of “before-after” is local scaffolding. We backfill causal arcs because compact narratives travel better through memory and conversation. This is useful. It also hides uncertainty. The after-the-fact confidence that a belief was inevitable—hindsight bias—is a compression artifact. The cure is not to abandon narrative; it’s to annotate it. Keep markers for uncertainty, explicit placeholders for alternative branches not taken. Think of these as side channels your metacognition maintains, preventing the main story from erasing valuable doubt.
Predictive-processing accounts sharpen the point. Perception is model-testing; action is model-correction. Precision (expected reliability) decides whether to rely on the model or renegotiate it via new evidence. Metacognitive signals—felt confidence, the urge to check, the discomfort of cognitive dissonance—are not moral failings. They are control variables surfacing the system’s precision estimates into conscious guidance. When that surfacing fails, odd things happen. Anxiety becomes perpetual sampling without stopping rules. Overconfidence becomes zero-sampling with brittle codes. Many spiritual traditions, in plainer language, have long trained the same skill: watch attention itself, label internal weather, widen the window in which updates can occur. No need to freight this with mysticism. It’s channel management.
There’s also an ethical undertow. Our slow “moral memory”—accumulated norms, taboos, scripts—is a culture-level compression of harms and goods discovered the hard way. It is lossy and biased, yet also robust in ways a single mind’s fast improvisation isn’t. Metacognition at the personal level should learn when to defer to this longer code and when to challenge it. The information lens doesn’t tell you what is right. It tells you when your private code is outgunned by a larger, older compressor—or when the old code is overfit to dead contexts and must be softened. That dilemma won’t resolve cleanly. It’s the point.
Practice and Friction: Classrooms, Clinics, and Codebases
Education first. Students don’t just need content; they need calibrated metacognitive control. One simple protocol: write down a prediction and a confidence percentage before answering a question; after feedback, compute the Brier score (a proper scoring rule) to align felt confidence with actual accuracy. Over weeks, this reduces both false certainty and wasteful hedging. Another: redesign notes as compressed codes. Force each lecture into a handful of atomic claims and proofs, then re-encode them in a second modality ( sketches, gestures, analogies ). Redundancy against noise. Finally, teach stop rules: when further reading has low expected information gain, switch to retrieval practice. Students learn to treat attention as a scarce channel, not an ethical failing. The effect is not only better grades. It’s lower anxiety, because the system now knows how it will decide to stop.
Clinical settings next. Many therapies already operate as information-control training whether they say so or not. Metacognitive Therapy (MCT) for anxiety targets “cognitive attentional syndrome”: worry loops that endlessly sample possibilities without updating global beliefs. Reframed informationally, worry is unbounded exploration with no cost accounting. The intervention adds prices: time limits, forced externalization (write once, do not rehash), and scheduled “worry windows” that cap bandwidth. For depression, behavioral activation functions as a data acquisition plan: alter priors by collecting discrepant evidence in the wild, not in the head. For obsessive-compulsive patterns, exposure with response prevention tampers with the system’s miscalibrated precision on threat cues; the metacognitive move is to tolerate high entropy without compulsive sampling. The shared skill: sensing when uncertainty belongs in the world (needs action) vs. in the model (needs update) vs. in the monitor (needs to stop monitoring).
Now code. In machine learning, we’ve started to bolt on introspective modules: calibration layers, uncertainty estimates, debate mechanisms, monitoring of chain-of-thought. The hope is mechanical metacognition. Some of it works. Better calibration reduces overconfident errors; active learning prioritizes high-uncertainty samples; ensemble disagreement approximates “second thoughts.” But there are hazards. If the monitor’s objective is too close to the main loss, the system will game its own doubt: performative hedging, confidence masking, or just fluent nonsense with a humility costume. Worse, a model without a slow, inherited moral memory will optimize local objectives aggressively, even if its “reflection” seems careful. You can’t patch this with a dashboard. You need constraints with teeth—constitutional training that actually bites during optimization, oversight that cannot be quietly routed around, open-sourced scrutiny resistant to incentive capture.
There is a workable engineering analog to culture’s long compression. Bake in norms as low-bandwidth but durable channels: rules that are cheap to check yet expensive to violate. Keep the “why” (the narrative) near the “what” (the rule) to avoid decontextualized drift. Track calibration over real tasks, not synthetic benchmarks. And accept that metacognitive loops—the system watching itself—carry costs. You pay in speed to buy reliability. Sometimes you pay and get theater. The art is to size the monitor to the channel: enough redundancy to survive noise; not so much that the code ossifies.
Across classroom, clinic, and codebase, one principle recurs: optimize the second wire. Not just how well you think, but how well you track the value of thinking. In a world read as pattern and constraint—information as the substrate rather than scenery—this isn’t decoration. It’s survival. And sometimes, yes, restraint. Not every surprise demands a theory; not every doubt deserves a meeting. The discipline is to know, moment by moment, which is which.
Sydney marine-life photographer running a studio in Dublin’s docklands. Casey covers coral genetics, Irish craft beer analytics, and Lightroom workflow tips. He kitesurfs in gale-force storms and shoots portraits of dolphins with an underwater drone.