← back to index

The Society of Mind

Marvin Minsky ยท 1986 ยท full text ยท MIT OCW lectures
See also: The Emotion Machine (2006), the sequel that replaces agents with layered critics and "ways to think."

Intelligence emerges from vast numbers of mindless agents organized into agencies. No single agent is intelligent. No single principle explains the whole. The power comes from diversity.

The argument

wpMinsky's book is 270 one-page essays organized into 30 chapters. The running example throughout is a child building with blocks โ€” a task that looks simple but requires dozens of cooperating processes. Visual parsing, spatial reasoning, motor control, goal management, error recovery.

The core claim: there is no central executive. The mind is a society of agents, each mindless, that compose into agencies. Agencies handle domains (locomotion, language, spatial reasoning) and compete for control through winner-take-all switching. No compromise.

BUILDER SEE GRASP PLACE agents: mindless, do one thing agencies: groups that achieve what no agent can alone a child building with blocks needs all three agencies at once
record K-line replay K-line: record the active set on success, replay on similar cue

Architectural primitives

Agents and agencies

An agent does one thing. It has no understanding of what it does or why. An agency is a group of agents that together achieve a function none could achieve alone. The mind is agencies all the way up. There is no top.

Proto-specialists are innate agencies that handle survival domains: locomotion, feeding, defense. They are the boot-loader. Everything else is built on top through learning.

Memory agents (the nemes)

Mechanism What it does Parts bin analogue
K-lineRecords which agents were active during a success. Reactivates that constellation on a similar cue.Episodic memory / case-based reasoning
PolynemeA permanent K-line. Activates partial states across many agencies simultaneously.Embedding vector / pub-sub broadcast
MicronemeA diffuse contextual signal. Sets mood or context globally rather than activating specific representations.Context injection / system prompt
PronomeShort-term memory pointer. Controls which role-slot is currently active.Register / pointer / pronoun
IsonomeSignals many agencies to perform the same operation in parallel.Broadcast instruction / SIMD
ParanomeLinks pronomes across representational realms. Enables cross-domain analogy.Cross-modal attention / transfer learning
Representational structures
Structure What it does
FrameA skeletal data structure with slots and default assumptions. Defaults hold until contradicted by evidence.
Frame-arrayMultiple frames for the same object from different viewpoints. Shared slots enable perspective-switching.
Trans-frameRepresents an event: origin, destination, cause, motive, instrument, time. The verb of internal language.
Story-frameChains of trans-frames. Narrative as data structure.
Control and error correction
Mechanism What it does
A-brain / B-brainA-brain handles world interaction. B-brain monitors A-brain for errors โ€” infinite loops, faulty reasoning. B-brain can veto.
CensorSuppresses the activity preceding an unproductive action. Preemptive negative expertise.
SuppressorInhibits the unproductive action itself. Reactive negative expertise.
Difference-engineIdentifies gap between current state and goal. Invokes an operator known to shrink the gap.
Cross-exclusionWinner-take-all switching between proto-specialists. One agency suppresses all others. Emotions are cascade effects of this switching.
Key principles
Principle Statement
Papert's PrincipleMental growth depends not on acquiring new skills, but on acquiring new ways to use existing knowledge.
Exception PrincipleEasier to extend old frames by adding exceptions than to replace them.
Recursion PrincipleThe same strategies apply at every level of decomposition.
Predestined LearningSufficient innate constraint makes certain learning outcomes nearly inevitable.
Chapter map
Ch Title Key introductions
1PrologueAgents, agencies, blocks-world
2โ€“5Wholes, Conflict, Self, IndividualityHeterarchies, no compromise, circular causality
6Insight and IntrospectionB-brains, self-knowledge is dangerous
7Problems and GoalsDifference-engines, local responsibility
8A Theory of MemoryK-lines, level-bands, societies of memories
10โ€“12Papert's Principle โ†’ Learning MeaningUniframes, the exception principle
16EmotionProto-specialists, cross-exclusion cascades
18โ€“20Reasoning โ†’ ContextPolynemes, micronemes, connection lines
21โ€“22Trans-Frames โ†’ ExpressionPronomes, trans-frames, isonomes
24โ€“26Frames โ†’ Language-FramesFrames, defaults, frame-arrays, story-frames
27Censors and JokesSuppressors, censors, negative expertise
29โ€“30Realms of Thought โ†’ Mental ModelsParanomes, cross-realm analogy, recursive self-models

The Emotion Machine (2006)

Twenty years later, Minsky replaced agents with critics. Society of Mind asks how mindless agents compose into intelligence. The Emotion Machine asks how the system decides which agents to run. The answer: layered critics, each monitoring the level below.

Level What it does Who has it
1. Instinctive reactionsHard-wired responses. Pain, reflex, startle.Every agent
2. Learned reactionsConditioned responses. Flinch before the blow lands.Soar (chunking), LLMs (weights)
3. Deliberative thinkingSearch over alternatives. Means-ends analysis.Soar (impasse), ReAct
4. Reflective thinkingMonitor your own deliberation. "Am I stuck?"Reflexion, RLHF
5. Self-reflective thinkingReason about your own reflective strategies.Nobody
6. Self-conscious thinkingModel yourself as others see you.Nobody

Each level can interrupt, override, or redirect the one beneath it. Emotions are what happens when a critic triggers a large-scale mode switch, shutting down some resources and activating others. Anger narrows attention. Grief suspends goals. Architectural events, not decorations on top of cognition.

The critic hierarchy is A-brain/B-brain scaled to six layers. Soar's impasse mechanism gets two levels (substate monitors the superstate). MAKER gets two (verifier monitors worker). RLHF gets two (reward model monitors policy). Nobody has built six. Minsky argued you need all six before the system can reason about its own reasoning about its own reasoning. Whether he was right depends on whether self-reflection decomposes into a finite stack or requires something else entirely.

Where these ideas landed

Minsky never wrote pseudocode. His contribution is design patterns. But every architecture in this collection reinvented his ideas under different names.

K-line replay is episodic memory. Soar stores a working-memory snapshot at every decision cycle, retrieves by partial cue matching. LLM agents do the same with natural-language episode streams (Generative Agents) and MemGPT's paged retrieval. Record what was active when something worked; replay it when the cue matches.

The A-brain acts; the B-brain watches for errors and vetoes. That's RLHF: a reward model judging the policy model's output. It's MAKER: verifier agents challenging worker outputs. Constitutional AI is a B-brain that reads a constitution. Minsky had the pattern in 1986.

Censors and suppressors are negative expertise: learned knowledge about what not to do. Soar's reject preferences fire before ranking. MemAgent learns what to forget. Alignment filters are suppressors. The censor is more interesting. It suppresses the thought preceding the bad action. Preemptive negative expertise. No modern system does this cleanly.

Cross-exclusion is softmax. One agency wins, all others suppressed. Minsky used it for emotions: proto-specialists competing for control of the whole system. Attention heads compete the same way, but between tokens rather than entire agencies.

Polyneme broadcast is the embedding vector. One signal activates partial states across many agencies at once. Multi-head attention does this with learned projections into subspaces. Minsky's version is simpler and more honest about what it is: pub-sub.

Neighbors