Continuous Distributions
Hey students! š Welcome to one of the most fascinating topics in A-level Further Mathematics - continuous distributions! In this lesson, we'll explore how probability works when dealing with continuous variables (like height, time, or temperature) rather than discrete ones. You'll discover the power of uniform, normal, and exponential distributions, learn transformation techniques, and understand how the Central Limit Theorem connects everything together. By the end, you'll be able to model real-world scenarios with confidence and solve complex probability problems! šÆ
Understanding Continuous Probability Distributions
Unlike discrete distributions where we deal with specific values (like rolling a 6 on a die), continuous distributions deal with ranges of values. Imagine measuring the height of students in your school - you could get 165.3 cm, 165.31 cm, or 165.314 cm. The possibilities are infinite within any range! š
A probability density function (PDF) describes how probability is distributed across all possible values. The key difference from discrete distributions is that we can't ask "what's the probability of exactly 165.3 cm?" because that probability is essentially zero. Instead, we ask "what's the probability of being between 165 cm and 166 cm?"
The fundamental properties of any PDF f(x) are:
- $f(x) \geq 0$ for all x (probabilities can't be negative)
- $\int_{-\infty}^{\infty} f(x) dx = 1$ (total probability equals 1)
- $P(a \leq X \leq b) = \int_a^b f(x) dx$ (probability is the area under the curve)
Think of it like this: if you're painting a fence and the PDF shows how much paint you use per meter, the total paint used (area under the curve) must equal exactly one can of paint! šØ
The Uniform Distribution
The uniform distribution is the simplest continuous distribution - it's completely flat! Every value within a given range has exactly the same probability density. It's like a perfectly fair spinner that can land anywhere between two points with equal likelihood.
For a uniform distribution on the interval [a, b], the PDF is:
$$f(x) = \begin{cases} \frac{1}{b-a} & \text{if } a \leq x \leq b \\ 0 & \text{otherwise} \end{cases}$$
The mean is $\mu = \frac{a+b}{2}$ (exactly in the middle, as you'd expect!) and the variance is $\sigma^2 = \frac{(b-a)^2}{12}$.
Real-world example: Bus arrival times! If buses are scheduled every 20 minutes and you arrive at a random time, your waiting time follows a uniform distribution between 0 and 20 minutes. The probability of waiting between 5 and 10 minutes is $\frac{10-5}{20-0} = 0.25$ or 25%. š
Random number generators in computers often use uniform distributions. When you press "shuffle" on your music playlist, the algorithm typically starts with uniform random numbers to ensure each song has an equal chance of being selected first!
The Exponential Distribution
The exponential distribution models the time between events in processes where events occur continuously and independently at a constant average rate. It's the continuous cousin of the geometric distribution! ā°
The PDF is: $f(x) = \lambda e^{-\lambda x}$ for $x \geq 0$, where $\lambda > 0$ is the rate parameter.
Key properties:
- Mean: $\mu = \frac{1}{\lambda}$
- Variance: $\sigma^2 = \frac{1}{\lambda^2}$
- Memoryless property: $P(X > s + t | X > s) = P(X > t)$
The memoryless property is fascinating! It means that if you've already waited 10 minutes for a bus, the probability of waiting another 5 minutes is the same as if you'd just arrived. The past doesn't affect the future - the distribution "forgets" what happened before.
Real-world applications:
- Radioactive decay: The time until an atom decays follows an exponential distribution. If carbon-14 has a half-life of 5,730 years, then $\lambda = \frac{\ln(2)}{5730} \approx 1.21 \times 10^{-4}$ per year.
- Customer service: Time between phone calls at a call center
- Equipment failure: Time until a light bulb burns out or a hard drive fails
In reliability engineering, if electronic components have an average lifespan of 1000 hours, the probability a component fails within the first 100 hours is $P(X \leq 100) = 1 - e^{-100/1000} \approx 0.095$ or about 9.5%.
The Normal Distribution
The normal distribution is the superstar of statistics! š Also called the Gaussian distribution (after Carl Friedrich Gauss), it appears everywhere in nature and forms the foundation of many statistical methods.
The PDF is: $f(x) = \frac{1}{\sigma\sqrt{2\pi}} e^{-\frac{(x-\mu)^2}{2\sigma^2}}$
where $\mu$ is the mean and $\sigma$ is the standard deviation.
Key properties:
- Bell-shaped and symmetric around the mean
- 68% of values lie within 1 standard deviation of the mean
- 95% lie within 2 standard deviations
- 99.7% lie within 3 standard deviations (the "68-95-99.7 rule")
Why is it so important? The normal distribution appears naturally in many situations:
- Human characteristics: Heights, weights, IQ scores, blood pressure
- Measurement errors: When you measure something repeatedly, errors typically follow a normal distribution
- Natural phenomena: Daily temperature variations, rainfall amounts
For example, adult male heights in the UK have a mean of approximately 175.3 cm with a standard deviation of 7.1 cm. This means about 68% of men are between 168.2 cm and 182.4 cm tall!
The standard normal distribution (Z-distribution) has $\mu = 0$ and $\sigma = 1$. Any normal distribution can be converted to standard form using: $Z = \frac{X - \mu}{\sigma}$
Transformation Methods
Sometimes we need to find the distribution of a transformed random variable. If $X$ has a known distribution and we want to find the distribution of $Y = g(X)$, we can use several methods:
Method 1: CDF Transformation
- Find $F_Y(y) = P(Y \leq y) = P(g(X) \leq y)$
- Solve for X in terms of y
- Express in terms of $F_X$
- Differentiate to get $f_Y(y)$
Method 2: Jacobian Method (for more complex transformations)
If $Y = g(X)$ where g is monotonic and differentiable:
$f_Y(y) = f_X(g^{-1}(y)) \left| \frac{d}{dy}g^{-1}(y) \right|$
Example: If $X \sim \text{Uniform}(0,1)$ and $Y = -\ln(X)$, then $Y$ follows an exponential distribution with $\lambda = 1$. This is actually how many computer programs generate exponential random numbers from uniform ones! š»
The Central Limit Theorem
The Central Limit Theorem (CLT) is one of the most powerful results in statistics! It explains why the normal distribution is so prevalent in nature. šÆ
Statement: If $X_1, X_2, ..., X_n$ are independent random variables with the same distribution (mean $\mu$ and variance $\sigma^2$), then as $n$ becomes large, the sum $S_n = X_1 + X_2 + ... + X_n$ approaches a normal distribution with:
- Mean: $n\mu$
- Variance: $n\sigma^2$
More commonly, we look at the sample mean $\bar{X} = \frac{S_n}{n}$, which approaches:
$$\bar{X} \sim N\left(\mu, \frac{\sigma^2}{n}\right)$$
The amazing part: This works regardless of the original distribution! Whether you start with uniform, exponential, or any other distribution, the average always becomes approximately normal for large n.
Real-world applications:
- Quality control: Manufacturing processes use CLT to set control limits
- Polling: Election polls rely on CLT to estimate population opinions from samples
- Finance: Stock price movements and portfolio returns
- Medicine: Clinical trial results and drug effectiveness studies
For example, if you roll a fair die (uniform on {1,2,3,4,5,6}) many times and calculate the average, that average will be approximately normal with mean 3.5, even though individual rolls are definitely not normal! With just 30 rolls, the approximation is already quite good.
Practical Applications and Problem-Solving
Let's see how these distributions work together in real scenarios:
Quality Control Example: A factory produces bolts with lengths that should be 5.0 cm. Due to manufacturing variations, lengths follow $N(5.0, 0.1^2)$. What percentage of bolts are between 4.8 and 5.2 cm?
Using standardization: $P(4.8 < X < 5.2) = P\left(\frac{4.8-5.0}{0.1} < Z < \frac{5.2-5.0}{0.1}\right) = P(-2 < Z < 2) \approx 0.95$ or 95%.
Reliability Engineering: Electronic components fail according to an exponential distribution with mean time to failure of 2000 hours. What's the probability a component lasts at least 3000 hours?
$P(X > 3000) = e^{-3000/2000} = e^{-1.5} \approx 0.223$ or about 22.3%.
Conclusion
Continuous distributions provide powerful tools for modeling real-world phenomena where variables can take any value within a range. The uniform distribution handles situations with equal likelihood across an interval, the exponential distribution models waiting times and decay processes, and the normal distribution describes natural variations in countless applications. Transformation methods allow us to work with more complex relationships between variables, while the Central Limit Theorem explains why normal distributions appear so frequently in practice. These concepts form the backbone of statistical inference, quality control, risk assessment, and scientific research. Understanding these distributions gives you the mathematical foundation to analyze uncertainty and make informed decisions in an unpredictable world! š
Study Notes
⢠Continuous PDF properties: $f(x) \geq 0$, $\int_{-\infty}^{\infty} f(x) dx = 1$, $P(a \leq X \leq b) = \int_a^b f(x) dx$
⢠Uniform Distribution U(a,b): $f(x) = \frac{1}{b-a}$ for $a \leq x \leq b$, $\mu = \frac{a+b}{2}$, $\sigma^2 = \frac{(b-a)^2}{12}$
⢠Exponential Distribution: $f(x) = \lambda e^{-\lambda x}$ for $x \geq 0$, $\mu = \frac{1}{\lambda}$, $\sigma^2 = \frac{1}{\lambda^2}$
⢠Memoryless Property: $P(X > s + t | X > s) = P(X > t)$ for exponential distributions
⢠Normal Distribution: $f(x) = \frac{1}{\sigma\sqrt{2\pi}} e^{-\frac{(x-\mu)^2}{2\sigma^2}}$
⢠68-95-99.7 Rule: 68% within 1Ļ, 95% within 2Ļ, 99.7% within 3Ļ of mean
⢠Standardization: $Z = \frac{X - \mu}{\sigma}$ converts any normal to standard normal
⢠CDF Transformation: Find $F_Y(y) = P(g(X) \leq y)$, then differentiate for PDF
⢠Central Limit Theorem: Sample means approach $N\left(\mu, \frac{\sigma^2}{n}\right)$ for large n, regardless of original distribution
⢠CLT Applications: Quality control, polling, finance, medical research - anywhere sample averages are used
