Mental models and ways of seeing

Mental Models

The invisible assumptions and stories we carry about how the world works

Also known as: Worldviews, Assumptions, Paradigms

THE IDEA

The map you forgot you were using

Everyone carries a set of assumptions about how the world works. These aren’t ideas you consciously chose. They’re deeper than that - baked into how you see, what you notice, and what you think is possible. They’re mental models, and they run in the background like an operating system you never installed but can’t stop using.

A mental model is a simplified internal representation of how something works. You have one for how traffic flows, one for how your boss makes decisions, one for how economies grow, one for how relationships function. Each one is a compression of reality - useful enough to get through the day, wrong enough to get you into trouble when the world doesn’t match your map.

The power of mental models lies in their invisibility. When your model matches reality well enough, it feels like you’re seeing the world directly. You’re not. You’re seeing your model of the world, and the model is always simpler, always older, and always shaped by the experiences that built it. The fish doesn’t see the water. You don’t see your mental models. And the ones you can’t see are the ones that have the most influence over your decisions.

IN PRACTICE

The lenses nobody notices they’re wearing

A manager believes that people are fundamentally lazy and need external motivation to do good work. This mental model shapes every decision: tight deadlines, constant check-ins, performance-linked bonuses, surveillance of output. The team responds by doing exactly what’s measured and nothing more - confirming the manager’s model. A different manager, with a model that people are fundamentally motivated by purpose and autonomy, designs a completely different environment and gets completely different behaviour from the same people. Both managers believe they’re seeing reality. Both are seeing their model.

A city planner operates with the mental model that traffic congestion is caused by insufficient road capacity. Every solution involves building more roads. A planner with a different model - that congestion is caused by car dependency, which is caused by land use patterns - proposes completely different solutions: mixed-use zoning, walkable neighbourhoods, public transport investment. Same city, same congestion, different models, different futures.

A person who grew up in a household where conflict meant danger carries a mental model that disagreement is threatening. In adult relationships, they avoid difficult conversations, interpret normal friction as signs of disaster, and accommodate until resentment builds. The model was perfectly functional for a child in a volatile home. It’s destructive for an adult in a healthy partnership. But it doesn’t feel like a model. It feels like reality.

WORKING WITH THIS

Making the invisible visible

The first step is accepting that you have mental models and that they’re not the same as truth. This sounds obvious but it’s not how most people operate. Most people defend their models as facts. “That’s just how it works.” “People are like that.” “The market will always…”

To surface a mental model, look at the decisions it produces. When you or your team consistently reaches the same kind of conclusion, ask: what would we have to believe about the world for that to seem like the right answer? The beliefs underneath the decision are the mental model.

Once a model is visible, you can test it. Ask: where does this model work well? Where does it break down? What would someone with a different model see that I’m missing? The goal isn’t to have no mental models - that’s impossible. The goal is to hold them lightly, to know they’re maps rather than territory, and to be willing to update them when reality sends a clear enough signal. The hardest part isn’t changing a mental model. It’s admitting that it’s a model at all.

THE INSIGHT

The line to remember

The most dangerous mental models aren’t the wrong ones. They’re the ones you don’t know you have - the assumptions so deep they feel like facts.

RECOGNITION

When this is in play

You’re bumping into mental models when different people look at the same situation and see completely different problems. When a team keeps reaching the same kind of solution regardless of the problem. When someone says “that’s obvious” about something that isn’t obvious to anyone else in the room. When a policy or strategy fails and the response is to try the same approach harder rather than questioning the assumptions behind it. When you feel certain about something you’ve never tested - that certainty is your mental model talking.

Connected concepts

System Blindness

System blindness is what happens when mental models can't see structure - defaulting to individual blame

Leverage Points

Changing mental models is one of the highest leverage points in a system - near the top of Meadows' hierarchy

Complexity vs Complication

The mental model of 'everything is a machine' is what makes people treat complex problems as merely complicated

Feedback loops

Mental models filter which feedback we notice and which we ignore - reinforcing what we already believe

Boundaries

Mental models determine where we draw boundaries - what's included in 'the problem' and what's invisible

Causal Loop Diagrams

Drawing a CLD surfaces mental models - it forces you to state your assumptions about what causes what

Systems Mapping

Systems mapping surfaces mental models by forcing people to draw what they think is happening

Map is Not the Territory

Every mental model is a map - a simplification that helps you navigate but never captures the full territory

Transformability

Transformation requires changing mental models - the deepest, hardest, and highest-leverage kind of change

Boundary Critique

Boundaries reflect mental models - how you frame the system reveals what you assume about what matters

Multiple Perspectives

Each perspective carries its own mental models - its own assumptions about what's important and why

Learning Organisation

Surfacing and testing mental models is one of Senge's five disciplines of a learning organisation

Double-Loop Learning

Double-loop learning is the process of questioning and revising mental models, not just actions

Sensemaking

Sensemaking builds and updates mental models - the process by which our map of reality gets revised

Bounded Rationality

Mental models are how bounded rationality works - we simplify the world into workable maps because processing everything is impossible

Motivated Reasoning

Motivated reasoning protects mental models from being updated - we filter evidence to confirm rather than challenge

Narrative Fallacy

Narratives build and reinforce mental models - the stories we tell shape what we expect to happen next

Hindsight Bias

Hindsight bias reinforces mental models by making us think our predictions are better than they are

Participatory Systems Mapping

Participatory mapping surfaces hidden mental models - the process of drawing together makes assumptions visible

thinking assumptions perception change