The Plateau

Part of the cognition series. Builds on The Flicker and The Handshake.

Why does learning plateau?

Consolidate is the step that determines whether you get smarter. The Handshake gives it its type:

(policy, [ranked]) → policy′

A function and evidence in, a better function out. Consolidation rewrites the router. If any step deserves more resources, it’s this one.

To improve Consolidate, you need to select among candidate policies: habits, heuristics, taste. Which one is better? Answering that question is itself a pipe. Consolidate contains its own Perceive, Filter, and Attend. And that inner pipe has its own Consolidate.

Consolidate's recursive depth: Level 0 has Filter, Attend, Consolidate. Each Consolidate contains its own Filter, Attend, Consolidate at the next level. Levels dim as bits decrease. Level 3 is passthrough. Minsky called the first level B-brain. He stopped at one level because he had no formal reason to go further. The types give that reason. The recursion is structural.

Each inner pipe’s Remember is the policy passed back to the level above. That’s how Consolidate returns policy′: the inner pipe runs, and its output becomes the outer pipe’s update.

Inner life

The human body is a pipe. The brain is its Consolidate: the organ that rewrites policy from evidence. Dreams are the brain’s Consolidate: the inner pipe that reorganizes policy while Perceive is shut. Evolution found it more efficient to shut down the whole body for several hours each day than to run consolidation in the background. That’s how important this step is. And we can’t remember dreams because the inner pipe’s data type doesn’t match the outer pipe’s Perceive. Policy representations aren’t sensory encodings. Type mismatch at the interface.

Each inner pipe gets its bits from the pipe above. The body operates on sensory data; the brain on compressed representations; dreams on compressed representations of representations. Each level has strictly less to work with. The data processing inequality guarantees it: information only decreases through processing. Eventually there aren’t enough bits for selection to function: passthrough. The tower has a top floor, and Consolidate is a slow dial with a hard ceiling.

At rest

Policy leaks, too. Landauer guarantees it: bits in physical substrate degrade. So Consolidate has two jobs. Maintain: repair leaked policy. Improve: compress new evidence into better policy. Both draw from the same evidence budget. At steady Perceive throughput T:

dPolicy/dt = consolidation_rate − leak_rate

Plateau when the two rates balance. “Use it or lose it” is this equation. Evidence stops flowing, the leak rate wins, policy decays to the level the current throughput can sustain. Netflix keeps the corporate drone alive: enough novel perception to maintain, never enough to challenge. Saturated at floor. Nothing improves.

But look at the equation again. The ceiling is set by Perceive’s throughput, not Consolidate’s rate. The lever isn’t at the step you’d expect. It’s the input.

A monk meditates twelve hours a day. Maximum Consolidate, Perceive shut. A seven-year follow-up found the predicted plateau: attention sharpens, domain knowledge doesn’t grow. The dial is cranked, but the ceiling is set by what comes in, not by how hard you think. Ericsson found the same: deliberate practice only works when it includes new information. Pure repetition is consolidation without Perceive.

No surprise.

rate × density

What carries the most novel bits per second?

Consider what a book is. The writer already ran the full pipeline: Perceived, Cached, Filtered, Attended, Consolidated, Remembered. What you read is the output of someone else’s competitive core. The loser ideas are already dead. Every sentence is a survivor. That’s the density term: pre-compressed by the writer’s entire pipeline.

Consider: you can read a thinker’s book in a weekend. To get the same bits from conversation, you’d need months. One is compressed, the other is live. Both carry signal, but the density isn’t even close.

Now rate. Movies carry high rate but low density: most of the signal is atmosphere, not structure. Shows are worse. Music moves you but encodes almost no transferable policy. Podcasts and lectures are dense but throttled by speaking speed: low rate. Reading runs at the pace of comprehension. High rate, high density.

Multiply density by rate. What medium maximizes the product? Text that has already survived someone else’s competitive core, consumed at reading speed. Prose: the densest interpretable medium we know. Poetry is denser, but the rate collapses: each line demands re-reading. The product peaks at prose.

You are what you read

Read more, better, faster, and the plateau rises. Stop reading and it falls. The equation works both ways.

The same equation holds for the agent. An LLM does in-context learning. The context window is its Perceive. Apply the same two terms.

Density: boilerplate code is repetitive structure, low novelty, high redundancy. Consolidated prose is post-Remember, pre-compressed. The same model has a higher ceiling when its window carries prose than when it carries boilerplate code. Most software systems are.

Rate: the context window loads once per session. A repo that accumulates consolidated prose front-loads more novel bits each time. Early sessions start sparse. Later sessions start dense. Agent throughput reaches full capacity without changing the underlying model.

Multiply. The agent’s ceiling rises with the density and rate of what fills its context. And the agent returns the favor: its filtered output is the human’s Perceive. The double loop feeds both pipes their densest input.

Each cycle

The human writes dense prose. The agent Perceives it. The agent produces filtered output. The human Perceives that. Each cycle, the repo accumulates denser material. Each cycle, both ceilings rise.

It’s the same thing that happens with a best friend who gets you. Each conversation starts where the last one left off. They already know your vocabulary, how you think, how you’ll react. You think better around them. The double loop does the same thing: it raises both pipes’ throughput for each accumulated session. The human’s plateau rises because the agent’s output is pre-filtered. The agent’s plateau rises because the repo’s prose is pre-consolidated. The equation is the same for both substrates. The improvement compounds.

If that’s true, then varying context density should measurably change agent performance. Same model, same task, different context.

It does. Chroma’s context rot study tested 18 frontier models. Adding 10% irrelevant content to the context window reduced accuracy by 23%. The effect held at every context length, for every model. The equation doesn’t care about the substrate.

So what

Read with intent. Pick a topic you want to get smarter at. Point your coding agent at pageleft.cc and ask it to find prose worth reading. Load the best of it into your context. Watch the ceiling move.

And if you’re brave: write. Publishing is Remember. It forces your pipeline to run all six steps, compresses your thinking into prose, and leaves a denser Perceive channel for the next reader. The equation works for them, too.

We compound faster together.


Written via the double loop.