Further Reading: Probability — The Foundation of Inference

Books (Start Here)

Mlodinow, L. (2008). The Drunkard's Walk: How Randomness Rules Our Lives. Vintage. This is the single best book for developing intuition about probability. Mlodinow — a physicist and screenwriter — covers the birthday problem, the Monty Hall problem, the law of large numbers, the gambler's fallacy, and dozens of other probability puzzles through vivid stories and accessible explanations. If you read one book alongside this chapter, make it this one. It covers the same conceptual territory but with far more historical depth and storytelling. The chapters on how courts misuse probability and how Wall Street underestimates risk are particularly compelling.

Olofsson, P. (2015). Probabilities: The Little Numbers That Rule Our Lives (2nd ed.). Wiley. A conversational, example-rich introduction to probability written for people who don't consider themselves math enthusiasts. Olofsson covers the classical, relative frequency, and subjective approaches (Section 8.2) with humor and clarity. His treatment of the Monty Hall problem (Case Study 1) includes a wonderful collection of the angry letters sent to Marilyn vos Savant. A gentle but rigorous companion to Chapter 8.

Silver, N. (2012). The Signal and the Noise: Why So Many Predictions Fail — but Some Don't. Penguin. Nate Silver — the election forecaster behind FiveThirtyEight — explores how probability is used (and misused) in weather forecasting, poker, earthquake prediction, baseball, and political polling. The chapter on weather forecasting directly addresses the question from Section 8.2: "What does a 30% chance of rain actually mean?" Silver's thesis — that better probabilistic thinking leads to better predictions — is a book-length argument for the importance of this chapter.

Ellenberg, J. (2014). How Not to Be Wrong: The Power of Mathematical Thinking. Penguin. Mathematician Jordan Ellenberg dedicates several chapters to probability, including an excellent treatment of expected value (a preview of Chapter 10), the law of large numbers, and why lottery winners' strategies are meaningless. His discussion of the "improbability principle" — why unlikely events happen all the time — is a great antidote to the surprise that students feel when learning about the birthday problem (Section 8.1).

Rosenthal, J. S. (2006). Struck by Lightning: The Curious World of Probabilities. Joseph Henry Press. A probability textbook in disguise — Rosenthal, a University of Toronto statistician, covers every major topic from Chapter 8 through Chapter 10 using accessible real-world examples. His chapters on gambling odds, coincidences, and genetics are excellent supplements to Sections 8.3, 8.6, and the sports case study. More mathematical than Mlodinow but still very readable.

Articles and Papers

Gilovich, T., Vallone, R., & Tversky, A. (1985). "The Hot Hand in Basketball: On the Misperception of Random Sequences." Cognitive Psychology, 17(3), 295-314. The landmark study demonstrating that basketball fans and players perceive "hot streaks" where statistical analysis finds none. This paper is referenced in Case Study 2 and provides the empirical foundation for the gambler's fallacy discussion in Section 8.3. It's readable and uses only basic probability — perfect for students at this level.

Miller, J. B., & Sanjurjo, A. (2018). "Surprised by the Hot Hand Fallacy? A Truth in the Law of Small Numbers." Econometrica, 86(6), 2019-2047. The paper that partially overturned Gilovich, Vallone, and Tversky (above) by identifying a subtle statistical bias in their methodology. Miller and Sanjurjo found weak evidence for the hot hand after correcting this bias. The paper is technically demanding, but the introduction and discussion sections are accessible and provide a beautiful example of how probability concepts can lead even experts astray.

vos Savant, M. (1990). "Game Show Problem." Parade Magazine, September 9, 1990. The original column where Marilyn vos Savant presented the Monty Hall problem and its solution. The column, and the thousands of letters it provoked, is a case study in how deeply people resist counterintuitive probability results. Archived versions are available online. Read it alongside Case Study 1 for the full experience.

Kahneman, D., & Tversky, A. (1972). "Subjective Probability: A Judgment of Representativeness." Cognitive Psychology, 3(3), 430-454. The foundational paper on how people actually think about probability — and how those intuitions systematically differ from the formal rules. Kahneman and Tversky showed that people estimate probabilities by how "representative" an outcome seems, leading to predictable biases. This connects directly to the gambler's fallacy (Section 8.3) and the birthday paradox (Section 8.1). Accessible and fascinating.

Videos

3Blue1Brown — "But What Are the Chances?" (Simulating the Birthday Problem) (YouTube, ~18 min) Grant Sanderson's stunning visual explanation of the birthday problem (Section 8.9) uses animated probability calculations to make the math feel intuitive. His visualization of how the probability climbs with each new person is the best visual supplement to this chapter. Watch this alongside Section 8.9.

StatQuest with Josh Starmer — "Probability Is Not Intuitive" (YouTube, ~12 min) Starmer covers the gambler's fallacy, conditional probability pitfalls, and why our brains consistently misjudge probabilities. His whiteboard-and-animation style makes the concepts from Sections 8.3-8.6 extremely clear. A good review resource before moving to Chapter 9.

Khan Academy — "Basic Probability" (khanacademy.org, ~2 hours) A complete video series covering the classical approach, sample spaces, the addition rule, the multiplication rule, and contingency tables. Sal Khan's step-by-step approach is ideal for students who want additional worked examples beyond those in the chapter. Free and self-paced.

Numberphile — "The Monty Hall Problem" (YouTube, ~8 min) A concise, well-produced explanation of the Monty Hall problem featuring a real mathematician working through the logic with physical doors and toy cars. The visual demonstration makes the 100-door version (Case Study 1) especially compelling.

Veritasium — "The Bayesian Trap" (YouTube, ~11 min) While primarily about Bayes' theorem (Chapter 9), this video starts with the foundational probability concepts from Chapter 8 — particularly false positives and the base rate. Watching it after finishing this chapter provides an excellent preview of where probability is heading next.

Interactive and Online Resources

Seeing Theory (seeing-theory.brown.edu) An interactive probability visualization tool created by Daniel Kunin at Brown University. Chapter 1 (Basic Probability) lets you flip coins, roll dice, and draw cards while watching the law of large numbers converge in real time. Chapter 2 (Compound Events) visualizes the addition and multiplication rules with animated Venn diagrams. This is the single best interactive supplement to Chapter 8 — bookmark it.

Khan Academy: Probability and Statistics (khanacademy.org) The probability unit covers everything from basic probability through Bayes' theorem, with practice problems and immediate feedback. Particularly strong on contingency tables and the addition rule. Free.

Setosa.io — "Conditional Probability" (setosa.io/conditional/) An interactive visualization of contingency tables, marginal probabilities, and joint probabilities. You can drag sliders to change table values and watch the probabilities update in real time. Excellent for building intuition about the relationship between cell counts and probabilities (Section 8.7). Previews Chapter 9.

The Monty Hall Simulator (multiple versions available online) Several websites let you play the Monty Hall game hundreds of times and track your win rate for switching vs. staying. Search for "Monty Hall simulator" and play at least 50 rounds with each strategy. Seeing the 2:1 ratio emerge from your own data is more convincing than any explanation.

Brilliant.org — "Introduction to Probability" (brilliant.org) Brilliant's problem-based approach to learning probability is excellent for students who want more practice with challenging puzzles. The basic course is free and covers the birthday problem, the Monty Hall problem, and the gambler's fallacy with interactive problems that provide immediate feedback.

Podcasts

Radiolab — "Stochasticity" (radiolab.org) A beautifully produced episode exploring randomness, probability, and the human struggle to accept uncertainty. The hosts interview a woman who won the lottery four times (improbable but not impossible — the law of large numbers across millions of players), discuss the gambler's fallacy with casino experts, and explore why our brains resist probabilistic thinking. About 60 minutes. One of the best episodes of any podcast on probability.

Freakonomics Radio — "How to Make a Bad Decision" (freakonomics.com) An episode exploring how people systematically misuse probability in decision-making — from medical testing to financial planning. The discussion of base rate neglect (a preview of Chapter 9's Bayes' theorem) builds directly on the probability foundations from this chapter.

Cautionary Tales with Tim Harford — "The Gambler Who Beat Roulette" (pushkin.fm) Tim Harford tells the story of Joseph Jagger, who hired assistants to record the outcomes of roulette wheels at Monte Carlo in 1873 — and found a biased wheel. This is the relative frequency approach to probability (Section 8.2) used in the real world to exploit a casino. The episode also covers the law of large numbers and why Jagger's strategy eventually stopped working.

Looking Ahead

The probability concepts from this chapter are the foundation for everything that follows:

  • Chapter 9 (Conditional Probability and Bayes' Theorem): What happens when new information changes probabilities? You'll learn P(A|B) — the probability of A given B — and Bayes' theorem, which combines prior beliefs with evidence. This directly extends the addition and multiplication rules from this chapter.

  • Chapter 10 (Probability Distributions and the Normal Curve): You'll move from discrete events (dice, coins) to continuous variables (heights, test scores), formalizing the connection between the relative frequency approach and the normal distribution. The probability rules from this chapter apply directly.

  • Chapter 11 (Sampling Distributions and the Central Limit Theorem): The law of large numbers (Section 8.3) reappears as the theoretical foundation for why sample means behave predictably. The CLT — the most important theorem in introductory statistics — is built on the probability framework you've just learned.

  • Chapter 13 (Hypothesis Testing): A p-value is a probability — specifically, the probability of observing data as extreme as yours, assuming the null hypothesis is true. Every concept from this chapter (sample spaces, events, complement rule) appears in hypothesis testing.

  • Chapter 19 (Chi-Square Tests): The contingency tables from Section 8.7 return for formal hypothesis testing. You'll test whether two categorical variables are independent using the same tables you've just learned to build.

  • Chapter 24 (Logistic Regression): Logistic regression predicts probabilities — specifically, the probability of a binary outcome (yes/no, pass/fail, buy/don't buy). The probability foundations from this chapter are literally the output of logistic regression models.