4. Statistics

Probability Theory

Axioms of probability, conditional probability, Bayes' theorem, independence and combinatorial methods for modelling uncertainty.

Probability Theory

Hey there students! 🎲 Welcome to one of the most fascinating areas of mathematics - probability theory! This lesson will take you on a journey through the mathematical framework that helps us understand uncertainty and make predictions about random events. By the end of this lesson, you'll master the fundamental axioms of probability, understand conditional probability and Bayes' theorem, grasp the concept of independence, and learn how combinatorial methods help us model uncertainty in the real world. Get ready to discover how probability shapes everything from weather forecasts to medical diagnoses! 🌟

The Foundation: Axioms of Probability

Let's start with the building blocks of probability theory - the axioms! Just like how geometry has its postulates, probability theory has three fundamental axioms that form its mathematical foundation.

Axiom 1: Non-negativity 📊

For any event A, the probability P(A) ≥ 0. This simply means probabilities can never be negative - it wouldn't make sense to have a -30% chance of rain!

Axiom 2: Normalization

The probability of the entire sample space (all possible outcomes) equals 1. Written mathematically: P(S) = 1, where S is the sample space. This makes perfect sense - something must happen, so the total probability is 100%!

Axiom 3: Countable Additivity

For mutually exclusive events A₁, A₂, A₃, ..., the probability of their union equals the sum of their individual probabilities: P(A₁ ∪ A₂ ∪ A₃ ∪ ...) = P(A₁) + P(A₂) + P(A₃) + ...

These axioms might seem simple, but they're incredibly powerful! From these three rules, we can derive all the other probability laws you'll encounter.

Let's see this in action with a real example. In the UK National Lottery, the probability of winning the jackpot is approximately 1 in 45,057,474, or P(jackpot) ≈ 0.0000022. The probability of not winning is P(no jackpot) = 1 - 0.0000022 ≈ 0.9999978. Notice how these probabilities are non-negative and sum to 1! 🎰

Conditional Probability: When Information Changes Everything

Now students, let's explore one of the most practical concepts in probability - conditional probability! This is where things get really interesting because it shows how new information can completely change our understanding of a situation.

Conditional probability is the probability of event A occurring given that event B has already occurred. We write this as P(A|B) and calculate it using:

$$P(A|B) = \frac{P(A \cap B)}{P(B)}$$

provided that P(B) > 0.

Here's a compelling real-world example from medical testing. Let's say you're taking a test for a rare disease that affects 1 in 1000 people. The test is 99% accurate for both detecting the disease when present and confirming its absence when not present.

If you test positive, what's the probability you actually have the disease? Many people would guess 99%, but let's calculate using conditional probability! 🔬

Let D = having the disease, T+ = testing positive

  • P(D) = 0.001 (disease prevalence)
  • P(T+|D) = 0.99 (test sensitivity)
  • P(T+|not D) = 0.01 (false positive rate)

Using the law of total probability:

P(T+) = P(T+|D) × P(D) + P(T+|not D) × P(not D)

P(T+) = 0.99 × 0.001 + 0.01 × 0.999 = 0.00099 + 0.00999 = 0.01098

Now, P(D|T+) = P(T+ ∩ D) / P(T+) = (0.99 × 0.001) / 0.01098 ≈ 0.09

Surprisingly, even with a positive test, there's only about a 9% chance you actually have the disease! This counterintuitive result highlights why understanding conditional probability is crucial in interpreting medical tests, legal evidence, and many other real-world scenarios.

Bayes' Theorem: The Ultimate Update Rule

Building on conditional probability, we arrive at Bayes' theorem - one of the most important results in probability theory! Named after Reverend Thomas Bayes, this theorem shows us how to update our beliefs when we receive new information.

Bayes' theorem states:

$$P(A|B) = \frac{P(B|A) \times P(A)}{P(B)}$$

This can also be written in its expanded form:

$$P(A|B) = \frac{P(B|A) \times P(A)}{P(B|A) \times P(A) + P(B|A^c) \times P(A^c)}$$

Let's see Bayes' theorem in action with a fascinating example from artificial intelligence! Modern spam filters use Bayesian methods to classify emails. Suppose we want to determine if an email containing the word "lottery" is spam.

From historical data:

  • P(spam) = 0.4 (40% of emails are spam)
  • P("lottery"|spam) = 0.8 (80% of spam emails contain "lottery")
  • P("lottery"|not spam) = 0.05 (5% of legitimate emails contain "lottery")

Using Bayes' theorem:

P(spam|"lottery") = (0.8 × 0.4) / (0.8 × 0.4 + 0.05 × 0.6) = 0.32 / (0.32 + 0.03) = 0.32 / 0.35 ≈ 0.91

So an email containing "lottery" has a 91% probability of being spam! This is exactly how your email provider protects you from unwanted messages. 📧

Independence: When Events Don't Influence Each Other

Independence is a crucial concept that students needs to master! Two events A and B are independent if the occurrence of one doesn't affect the probability of the other occurring.

Mathematically, events A and B are independent if and only if:

P(A ∩ B) = P(A) × P(B)

Alternatively, we can check independence using conditional probability:

P(A|B) = P(A) and P(B|A) = P(B)

A classic example of independence comes from coin flipping. If you flip a fair coin twice, the result of the first flip doesn't influence the second flip. P(heads on flip 1) = 0.5, P(heads on flip 2) = 0.5, and P(heads on both flips) = 0.5 × 0.5 = 0.25.

However, be careful not to confuse independence with mutually exclusive events! If two events are mutually exclusive (can't both occur), they're actually dependent because knowing one occurred tells us the other definitely didn't occur.

Here's a real-world application: In quality control at manufacturing plants, companies assume that defects in different products are independent events. If the defect rate is 2% per item, then the probability that two randomly selected items are both defective is 0.02 × 0.02 = 0.0004 or 0.04%. This assumption helps companies predict quality outcomes and set appropriate testing protocols. 🏭

Combinatorial Methods: Counting Our Way to Probability

Finally students, let's explore how combinatorial methods help us calculate probabilities, especially when dealing with equally likely outcomes! When all outcomes in a sample space are equally likely, we can calculate probability using:

$$P(A) = \frac{\text{Number of favorable outcomes}}{\text{Total number of possible outcomes}}$$

The key combinatorial tools you'll need are:

Permutations: When order matters, the number of ways to arrange r objects from n objects is:

$$P(n,r) = \frac{n!}{(n-r)!}$$

Combinations: When order doesn't matter, the number of ways to choose r objects from n objects is:

$$C(n,r) = \binom{n}{r} = \frac{n!}{r!(n-r)!}$$

Let's apply this to a real scenario! In the UK EuroMillions lottery, players choose 5 numbers from 1 to 50, plus 2 "Lucky Star" numbers from 1 to 12. The probability of winning the jackpot is:

P(jackpot) = 1 / [C(50,5) × C(12,2)]

$= 1 / [2,118,760 × 66]$

$= 1 / 139,838,160$

≈ 0.0000007%

This explains why lottery jackpots grow so large - the odds are astronomically small! 🎲

Combinatorial methods are also essential in genetics. For example, if both parents carry one dominant (A) and one recessive (a) allele for a trait, the probability their child inherits the recessive phenotype (aa) is 1/4, calculated by considering all possible combinations: AA, Aa, aA, aa.

Conclusion

Throughout this lesson students, we've built a solid foundation in probability theory! We started with the three fundamental axioms that govern all probability calculations, then explored how conditional probability and Bayes' theorem help us update our understanding when new information becomes available. We discovered how independence allows us to multiply probabilities and learned to distinguish it from mutual exclusivity. Finally, we saw how combinatorial methods provide the counting tools necessary for calculating probabilities in complex scenarios. These concepts work together to form a powerful mathematical framework for understanding uncertainty, making predictions, and making informed decisions in everything from medical diagnosis to financial planning to artificial intelligence! 🎯

Study Notes

• Three Axioms of Probability: P(A) ≥ 0, P(S) = 1, P(A₁ ∪ A₂ ∪ ...) = P(A₁) + P(A₂) + ... for mutually exclusive events

• Conditional Probability Formula: P(A|B) = P(A ∩ B) / P(B), where P(B) > 0

• Bayes' Theorem: P(A|B) = [P(B|A) × P(A)] / P(B)

• Independence Condition: Events A and B are independent if P(A ∩ B) = P(A) × P(B)

• Law of Total Probability: P(A) = P(A|B₁)P(B₁) + P(A|B₂)P(B₂) + ... for partition B₁, B₂, ...

• Permutations Formula: P(n,r) = n! / (n-r)! when order matters

• Combinations Formula: C(n,r) = n! / [r!(n-r)!] when order doesn't matter

• Classical Probability: P(A) = (favorable outcomes) / (total possible outcomes) for equally likely outcomes

• Complement Rule: P(A') = 1 - P(A)

• Addition Rule: P(A ∪ B) = P(A) + P(B) - P(A ∩ B)

• Multiplication Rule: P(A ∩ B) = P(A) × P(B|A) = P(B) × P(A|B)

• Independence vs Mutual Exclusivity: Independent events can occur together; mutually exclusive events cannot

Practice Quiz

5 questions to test your understanding

Probability Theory — A-Level Further Mathematics | A-Warded