4. Statistics and Probability

Discrete Probability Distributions

Discrete Probability Distributions

Welcome, students! 🎯 In this lesson, you will explore how discrete probability distributions describe situations where the number of outcomes is countable, like the number of heads in coin tosses or the number of defective items in a box. These ideas are important in statistics because they help us model real-life events, make predictions, and understand randomness in a structured way.

Learning objectives

By the end of this lesson, you should be able to:

  • Explain the main ideas and terminology behind discrete probability distributions.
  • Use the rules of probability to solve problems involving discrete random variables.
  • Find and interpret the expected value and variance of a discrete distribution.
  • Connect discrete probability distributions to wider topics in statistics and probability.
  • Recognize common IB-style question patterns and apply correct reasoning.

What is a discrete probability distribution?

A discrete probability distribution describes the probabilities of all possible values of a discrete random variable. A discrete random variable is a variable that can take only countable values, usually whole numbers. For example, the number of students absent in a class on a given day could be $0, 1, 2, 3,$ and so on. It cannot be $2.4 students, so this is discrete.

Let $X$ be a discrete random variable. A probability distribution assigns a probability to each possible value of $X$. These probabilities must satisfy two key rules:

$$P(X=x) \ge 0$$

for every possible value $x$, and

$$\sum P(X=x) = 1.$$

This means all probabilities must be non-negative, and together they must add up to exactly $1$.

A simple example is the number of heads when tossing a fair coin twice. The possible values of $X$ are $0, 1,$ and $2. The probabilities are:

$$P(X=0)=\frac14, \quad P(X=1)=\frac12, \quad P(X=2)=\frac14.$$

These values form a discrete probability distribution because they are countable and the probabilities sum to $1$.

Key terms and notation

When working with discrete probability distributions, it helps to know the standard language.

  • Random variable: a numerical result of a random experiment.
  • Discrete: having countable values.
  • Probability distribution: a table, graph, or formula showing all possible values and their probabilities.
  • Expected value: the long-run average value of the random variable.
  • Variance: a measure of how spread out the values are.
  • Standard deviation: the square root of the variance.

If $X$ is a discrete random variable, then its expected value is written as $E(X)$ and is found using

$$E(X)=\sum xP(X=x).$$

This means you multiply each value of $x$ by its probability and add the results.

The variance is

$$\operatorname{Var}(X)=E(X^2)-[E(X)]^2.$$

The standard deviation is

$$\sigma=\sqrt{\operatorname{Var}(X)}.$$

These formulas are central in IB Mathematics Analysis and Approaches SL because they help you summarize a distribution numerically.

Building a discrete probability distribution

To create a discrete probability distribution, follow these steps:

  1. Identify the random variable $X$.
  2. List every possible value of $X$.
  3. Assign a probability to each value.
  4. Check that the probabilities are all between $0$ and $1$.
  5. Check that the total probability is $1$.

Example: tossing two fair coins

Let $X$ be the number of heads when two fair coins are tossed.

The possible values are $0, 1,$ and $2.

  • $P(X=0)=\frac14$ because only $TT$ gives zero heads.
  • $P(X=1)=\frac12$ because $HT$ and $TH$ each give one head.
  • $P(X=2)=\frac14$ because only $HH$ gives two heads.

So the distribution is:

$$\begin{array}{c|ccc}

x & 0 & 1 & 2 \\hline

P(X=x) & $\frac14$ & $\frac12$ & $\frac14$

$\end{array}$$$

This table is a clean way to show a discrete probability distribution.

Finding expected value and spread

The expected value does not mean the result will happen every time. It means the average result over many repeated trials. Think of a game where you roll a die many times. You may not get the average result on one roll, but over many rolls the average becomes meaningful 🎲.

Using the coin example, the expected value is:

$$E(X)=0\left(\frac14\right)+1\left(\frac12\right)+2\left(\frac14\right).$$

So,

$$E(X)=0+\frac12+\frac12=1.$$

This means that in the long run, the average number of heads per two-coin toss experiment is $1$.

Now find $E(X^2)$:

$$E(X^2)=0^2\left(\frac14\right)+1^2\left(\frac12\right)+2^2\left(\frac14\right).$$

Thus,

$$E(X^2)=0+\frac12+1=\frac32.$$

Then the variance is

$$\operatorname{Var}(X)=\frac32-1^2=\frac12.$$

And the standard deviation is

$$\sigma=\sqrt{\frac12}.$$

The variance and standard deviation tell us how spread out the distribution is around the mean. A larger standard deviation means the values are more spread out.

Common distributions in IB reasoning

In IB Mathematics Analysis and Approaches SL, you may meet several familiar discrete distributions.

Binomial distribution

The binomial distribution is one of the most important discrete probability distributions. It applies when:

  • there are a fixed number of trials $n$,
  • each trial has only two outcomes, often called success and failure,
  • the probability of success is constant, $p$,
  • the trials are independent.

If $X\sim \mathrm{Bin}(n,p)$, then the probability of exactly $k$ successes is

$$P(X=k)=\binom{n}{k}p^k(1-p)^{n-k}.$$

For example, if a student guesses on $5$ true-or-false questions, each with probability $\frac12$ of being correct, then $X$ = number correct may be modeled by a binomial distribution with $n=5$ and $p=\frac12$.

Example calculation

Suppose $X\sim \mathrm{Bin}(4,0.3)$. Then the probability of exactly two successes is

$$P(X=2)=\binom42(0.3)^2(0.7)^2.$$

First calculate $\binom42=6$, so

$$P(X=2)=6(0.09)(0.49)=0.2646.$$

This kind of calculation is typical in IB questions.

The mean and variance of a binomial distribution are:

$$E(X)=np$$

and

$$\operatorname{Var}(X)=np(1-p).$$

These formulas are useful because they save time and show how the parameters affect the shape of the distribution.

Interpreting results in context

Statistics is not just about calculating numbers; it is about interpreting them correctly. If students sees $E(X)=3.2$ in a context, that does not mean $3.2$ will happen in one trial. It means that over many repeated trials, the average value is about $3.2$.

For example, if a factory finds that the expected number of defective items per box is $0.4$, that does not mean every box has exactly $0.4$ defects. It means that across many boxes, the average number of defects is $0.4$. Some boxes may have none, some may have one, and some may have more.

This is why discrete probability distributions are useful in real life. They help in quality control, sports statistics, gaming, insurance, and decision-making. For example, a sports analyst might model the number of goals a team scores in a match, while an insurance company might model the number of claims made in a month.

How this fits into Statistics and Probability

Discrete probability distributions connect directly to other parts of the statistics and probability syllabus.

  • In data collection and statistical description, we use data from samples to estimate patterns and compare to probability models.
  • In correlation and regression, we study relationships between variables; probability distributions help us understand uncertainty in data and predictions.
  • In conditional probability, we look at how probabilities change when information is given.
  • In continuous probability distributions, the variable can take any value in an interval, unlike the countable values in discrete distributions.

Discrete distributions are a bridge between counting outcomes and making predictions. They help students understand how probability models connect to real data and why randomness can still be studied with mathematics.

Conclusion

Discrete probability distributions describe situations where outcomes are countable and each outcome has a probability. The key ideas are that probabilities must be between $0$ and $1$, and their sum must equal $1$. You should also know how to calculate the expected value, variance, and standard deviation of a discrete random variable. The binomial distribution is especially important in IB Mathematics Analysis and Approaches SL because it models repeated independent trials with two outcomes.

When you understand discrete probability distributions, you are better prepared to analyze uncertainty, solve exam questions, and interpret results in context. Keep practicing with tables, formulas, and real-life examples, students, and these ideas will become much easier to use ✅

Study Notes

  • A discrete random variable takes countable values such as $0,1,2,\dots$.
  • A discrete probability distribution lists every possible value of $X$ and its probability.
  • Each probability must satisfy $P(X=x)\ge 0$.
  • All probabilities must add up to $1$, so $\sum P(X=x)=1$.
  • The expected value is $E(X)=\sum xP(X=x)$.
  • The variance is $\operatorname{Var}(X)=E(X^2)-[E(X)]^2$.
  • The standard deviation is $\sigma=\sqrt{\operatorname{Var}(X)}$.
  • The binomial distribution is written as $X\sim \mathrm{Bin}(n,p)$.
  • For a binomial distribution, $P(X=k)=\binom{n}{k}p^k(1-p)^{n-k}$.
  • The mean and variance of $X\sim \mathrm{Bin}(n,p)$ are $E(X)=np$ and $\operatorname{Var}(X)=np(1-p)$.
  • Discrete distributions are useful for modeling counts in real-life situations like defects, goals, heads, or survey responses.
  • They form an important part of Statistics and Probability because they help describe randomness clearly and mathematically.

Practice Quiz

5 questions to test your understanding