Mental Models

The thinking tools that the world's best decision-makers use to understand reality, solve problems, and avoid predictable errors.

Updated April 2026

What Is a Mental Model?

A mental model is a compressed representation of how something works. It's a thinking shortcut — not a lazy one, but a powerful one. Just as a map helps you navigate a city without walking every street, a mental model helps you navigate complex situations without having to reason from scratch every time.

The key insight, championed by Charlie Munger, is that no single model is sufficient. Reality is multifaceted. You need a latticework of mental models drawn from multiple disciplines — physics, biology, psychology, economics, engineering, history — and the skill to apply the right model to each situation.

The Man with a Hammer

"To a man with only a hammer, everything looks like a nail." If you only know one model — say, supply and demand — you'll try to apply it to everything, including situations where it doesn't fit. The cure is to build a toolkit of models so large that you always have the right tool available.

Foundational Models

Start here. These are the models that underpin everything else.

First Principles Thinking

What: Break a problem down to its fundamental truths — the things that are true regardless of convention, history, or assumption — and reason upward from there.

Example: Elon Musk on battery costs. The conventional wisdom was "batteries are expensive." First principles: What are batteries made of? Cobalt, nickel, lithium, carbon, a steel can, and some polymers. What do those raw materials cost on the commodity market? About $80/kWh. So why do batteries cost $600/kWh? Because of the existing manufacturing process, not the physics. The physics allows for much cheaper batteries — which is exactly what happened.

When to use: When you're told "that's just how it's done," when costs seem unjustifiably high, when you need genuine innovation rather than incremental improvement.

Danger: First principles thinking is slow and cognitively expensive. You can't reason from first principles about everything — you'd never get anything done. Use it selectively on high-stakes problems.

Inversion

What: Instead of asking "how do I achieve X?", ask "what would guarantee I fail at X?" Then systematically avoid those things.

Example: Want a great marriage? Instead of asking "how do I build a great marriage?", ask "how would I guarantee a terrible marriage?" Answer: stop communicating, take your partner for granted, never compromise, prioritise work over everything, be contemptuous. Now avoid those things. You won't have a perfect marriage, but you'll avoid the most common failure modes.

When to use: When the path to success is unclear but the path to failure is obvious. When you're planning a project and want to identify risks. In pre-mortems.

Second-Order Thinking

What: Consider not just the immediate consequence of an action, but the consequences of the consequences. First-order thinking asks "what happens next?" Second-order thinking asks "and then what?"

Example: A company cuts prices to gain market share (first-order: more customers). But competitors match the price cut (second-order: no market share gain). Now everyone has lower margins (third-order: reduced ability to invest in quality). The market commoditises (fourth-order: everyone loses).

When to use: Any decision with delayed or indirect consequences. Policy decisions. Competitive strategy. Hiring. Organisational changes.

The Map Is Not the Territory

What: Every model, plan, and mental representation is a simplification of reality. It captures some features and ignores others. The moment you forget this — the moment you confuse your model for reality — you stop seeing what's actually happening.

Example: Financial models predicted the 2008 housing market was safe. The models were based on historical data that assumed housing prices never decline nationally. The map (the model) was not the territory (reality). When the territory diverged from the map, people who trusted the map suffered catastrophic losses.

Decision-Relevant Models

Occam's Razor

What: When you have competing explanations, prefer the simplest one that adequately explains the data. Don't multiply entities unnecessarily.

Application: Your website traffic dropped. Before concluding that Google changed its algorithm to target you specifically, check whether your server went down, your SSL certificate expired, or you accidentally blocked crawlers in robots.txt.

Hanlon's Razor

What: Never attribute to malice that which can be adequately explained by incompetence, ignorance, or accident.

Application: Your colleague didn't respond to your email. Before assuming they're ignoring you, consider: they're overwhelmed, it went to spam, they forgot, or they thought it was informational and didn't need a reply. Most interpersonal friction is caused by carelessness, not ill intent.

Probabilistic Thinking

What: Think in probabilities, not certainties. Replace "this will happen" with "there's a 70% chance of this and a 30% chance of that." This forces you to consider multiple outcomes and prepare accordingly.

Application: Instead of "our product launch will succeed," think "there's a 60% chance of moderate success, a 25% chance of exceeding expectations, and a 15% chance of underperforming. What do we do in each scenario?"

Circle of Competence

What: Know the boundaries of what you actually understand well, and operate primarily within them. When you must operate outside them, seek expertise or proceed with extreme caution.

Application: Warren Buffett didn't invest in technology companies for decades — not because tech was bad, but because it was outside his circle of competence. He knew what he didn't know. Most people overestimate the size of their circle.

ModelCore PrincipleBest Applied When
First PrinciplesReason from fundamentals, not analogiesInnovation, cost reduction, challenging assumptions
InversionAvoid failure instead of seeking successRisk assessment, planning, relationship building
Second-Order ThinkingConsider downstream consequencesStrategy, policy, competitive moves
Map ≠ TerritoryModels are simplifications, not realityFinancial planning, forecasting, any model-based decision
Occam's RazorPrefer simpler explanationsDiagnosis, troubleshooting, hypothesis selection
Hanlon's RazorDon't assume maliceInterpersonal conflict, management, communication
Probabilistic ThinkingThink in likelihoods, not certaintiesForecasting, risk, investment, project planning
Circle of CompetenceKnow what you know and what you don'tCareer decisions, investment, delegation

Models from Psychology

Confirmation Bias

You seek information that confirms what you already believe and ignore information that contradicts it. This is the most dangerous cognitive bias because it's invisible — you feel like you're being objective when you're actually building a case for your existing view.

Counter: Actively seek disconfirming evidence. Ask "what would change my mind?" before forming a conclusion. Appoint a devil's advocate in team discussions.

Survivorship Bias

You study the winners and conclude that their traits caused their success, ignoring the losers who had the same traits. "College dropouts like Bill Gates and Mark Zuckerberg became billionaires, so maybe I should drop out too" — ignoring the millions of dropouts who didn't become billionaires.

Counter: Always ask "what does the data on failures look like?" Study the base rate, not just the highlights.

Loss Aversion

Losses feel roughly twice as painful as equivalent gains feel good. Losing £100 feels worse than finding £100 feels good. This makes people irrationally risk-averse — clinging to bad investments, staying in bad jobs, and avoiding changes that have positive expected value.

Counter: Evaluate decisions based on expected value, not on how the downside makes you feel. Ask "if I didn't already have this, would I choose it today?"

The Dunning-Kruger Effect

People with low competence in a domain overestimate their ability, while people with high competence underestimate theirs. The less you know, the less you know about how much you don't know.

Counter: Seek feedback from people who are better than you. Be suspicious of your confidence, especially in new domains. The feeling of certainty is not evidence of correctness.

Models from Systems Thinking

Feedback Loops

Positive (reinforcing): Output amplifies the input. Success breeds more success. Panic causes more panic. Compound interest. Viral growth. These create exponential change — for better or worse.

Negative (balancing): Output dampens the input. Thermostat keeps temperature stable. Competition reduces abnormal profits. These create stability and equilibrium.

Application: When something is growing or shrinking rapidly, look for the feedback loop driving it. To change the system, intervene in the loop — not in the symptom.

Leverage Points

Not all interventions are equal. Some changes have outsized effects because they sit at leverage points in the system. Changing a single rule, incentive, or information flow can transform outcomes more than massive effort applied in the wrong place.

Emergence

Complex behaviour arises from simple rules applied to many agents. No individual ant "plans" the colony. No individual trader "creates" market prices. The behaviour of the system cannot be predicted from the behaviour of its parts. This means some outcomes can't be designed — they can only be cultivated.

How to Build Your Latticework

Don't try to memorise models like flash cards. Instead: (1) Learn a model. (2) Look for it in the real world — news, work, relationships. (3) Apply it to a current problem. (4) Discuss it with someone. (5) Write about it. The goal isn't to recite definitions; it's to develop the instinct to recognise which model fits which situation. That instinct comes from practice, not study.