4. Continuous Distributions

Normal Approximation

Introduce the normal distribution, standardization (z-scores), and use normal approximation for sums and binomial cases.

Normal Approximation

Hey there students! šŸ‘‹ Today we're diving into one of the most powerful tools in statistics: the normal approximation. This lesson will help you understand how the bell-shaped normal distribution can be used to solve complex probability problems, especially when dealing with large datasets. By the end of this lesson, you'll master standardization using z-scores and learn how to use normal approximation for binomial distributions and sums of random variables. Get ready to unlock a statistical superpower that makes difficult calculations much easier! šŸš€

Understanding the Normal Distribution

The normal distribution, also known as the Gaussian distribution or bell curve, is the most important probability distribution in statistics. Picture a perfectly symmetrical bell shape - that's what we're working with! šŸ””

This distribution appears everywhere in nature and human behavior. For example, if you measured the heights of all high school students in your state, you'd find that most students cluster around an average height, with fewer students being very tall or very short. The same pattern emerges with test scores, reaction times, and even the weights of apples in an orchard.

The normal distribution has some amazing properties that make it incredibly useful:

Key Characteristics:

  • It's perfectly symmetrical around its center (the mean)
  • About 68% of all data falls within one standard deviation of the mean
  • About 95% falls within two standard deviations
  • About 99.7% falls within three standard deviations

This is called the 68-95-99.7 rule or the empirical rule. Think of it like this: if your class average on a test is 75 with a standard deviation of 10, then roughly 68% of students scored between 65 and 85, and 95% scored between 55 and 95.

The mathematical formula for the normal distribution is:

$$f(x) = \frac{1}{\sigma\sqrt{2\pi}} e^{-\frac{1}{2}\left(\frac{x-\mu}{\sigma}\right)^2}$$

Don't worry about memorizing this complex formula - what matters is understanding that $\mu$ represents the mean and $\sigma$ represents the standard deviation.

Standardization and Z-Scores

Here's where things get really cool! šŸ˜Ž Since there are infinite possible normal distributions (different means and standard deviations), statisticians created a way to standardize them all. This process is called standardization, and it uses something called z-scores.

A z-score tells you how many standard deviations away from the mean a particular value is. The formula is surprisingly simple:

$$z = \frac{x - \mu}{\sigma}$$

Where:

  • $x$ is your data point
  • $\mu$ is the mean of the distribution
  • $\sigma$ is the standard deviation

Let's say your friend scored 85 on a test where the class average was 75 and the standard deviation was 8. Their z-score would be:

$$z = \frac{85 - 75}{8} = \frac{10}{8} = 1.25$$

This means your friend scored 1.25 standard deviations above the average - pretty good! šŸ“ˆ

The Standard Normal Distribution is what we get when we standardize any normal distribution. It always has a mean of 0 and a standard deviation of 1. This is incredibly powerful because we can use standard normal tables (z-tables) to find probabilities for any normal distribution after standardizing.

For example, if you want to know what percentage of students scored below your friend's 85, you'd look up z = 1.25 in a standard normal table and find that about 89.4% of students scored lower. This means your friend performed better than nearly 9 out of 10 classmates!

Normal Approximation to the Binomial Distribution

Now for one of the most practical applications of normal approximation! šŸŽÆ When dealing with binomial distributions (think coin flips, true/false questions, or any yes/no scenarios), calculating probabilities can become incredibly tedious when the number of trials gets large.

Imagine trying to calculate the probability of getting exactly 45 heads in 100 coin flips using the binomial formula - you'd be there all day! This is where normal approximation saves the day.

When can we use normal approximation for binomial distributions?

The rule of thumb is when both $np \geq 5$ and $n(1-p) \geq 5$, where:

  • $n$ is the number of trials
  • $p$ is the probability of success

For a binomial distribution with parameters $n$ and $p$, we can approximate it with a normal distribution having:

  • Mean: $\mu = np$
  • Standard deviation: $\sigma = \sqrt{np(1-p)}$

Let's work through a real example: A pharmaceutical company claims their new medication is effective for 70% of patients. In a study of 100 patients, what's the probability that between 65 and 75 patients will show improvement?

First, check if we can use normal approximation:

  • $np = 100 \times 0.7 = 70 \geq 5$ āœ“
  • $n(1-p) = 100 \times 0.3 = 30 \geq 5$ āœ“

Great! Now we can approximate with a normal distribution:

  • $\mu = 70$
  • $\sigma = \sqrt{100 \times 0.7 \times 0.3} = \sqrt{21} \approx 4.58$

To find P(65 ≤ X ≤ 75), we standardize:

  • For X = 65: $z_1 = \frac{65 - 70}{4.58} \approx -1.09$
  • For X = 75: $z_2 = \frac{75 - 70}{4.58} \approx 1.09$

Using standard normal tables, P(-1.09 ≤ Z ≤ 1.09) ā‰ˆ 0.724, or about 72.4%.

Normal Approximation for Sums of Random Variables

The Central Limit Theorem is one of the most important concepts in statistics, and it's the foundation for using normal approximation with sums. 🌟 This theorem states that when you add up many independent random variables, their sum approaches a normal distribution, regardless of what the original distributions looked like!

This is absolutely mind-blowing when you think about it. You could be adding variables from completely different distributions - some uniform, some exponential, some weird and asymmetric - but their sum will still be approximately normal if you have enough of them.

Key points about the Central Limit Theorem:

  • Works best with sample sizes of 30 or more
  • The approximation gets better as sample size increases
  • The mean of the sum equals the sum of the individual means
  • The variance of the sum equals the sum of the individual variances (for independent variables)

Real-world example: Imagine you're managing a busy coffee shop and want to predict your daily revenue. Each customer's purchase is a random variable - some buy just coffee ($3), others get elaborate drinks and pastries ($15), and everything in between. Even though individual purchases vary wildly, your daily total (the sum of hundreds of purchases) will follow a predictable normal pattern!

If your average customer spends $7 with a standard deviation of $4, and you serve 200 customers daily:

  • Mean daily revenue: $\mu = 200 \times 7 = \$1,400
  • Standard deviation: $\sigma = \sqrt{200} \times 4 = \sqrt{200} \times 4 \approx \$56.57

You can now calculate the probability of earning between $1,300 and $1,500 in a day by standardizing and using normal tables!

Conclusion

Normal approximation is like having a statistical Swiss Army knife! šŸ”§ We've explored how the elegant bell curve of the normal distribution appears throughout nature and human behavior, learned to standardize any normal distribution using z-scores, and discovered how to use normal approximation to simplify complex binomial calculations and sums of random variables. The Central Limit Theorem shows us that even chaotic, random processes tend toward predictable patterns when we look at them in aggregate. These tools will serve you well in advanced statistics, research, and real-world problem-solving where exact calculations would be impractical or impossible.

Study Notes

• Normal Distribution: Bell-shaped, symmetric distribution characterized by mean (μ) and standard deviation (σ)

• 68-95-99.7 Rule: 68% of data within 1σ, 95% within 2σ, 99.7% within 3σ of the mean

• Z-score Formula: $z = \frac{x - \mu}{\sigma}$ (measures standard deviations from mean)

• Standard Normal Distribution: Normal distribution with μ = 0 and σ = 1

• Binomial Normal Approximation Conditions: Use when $np \geq 5$ and $n(1-p) \geq 5$

• Binomial Approximation Parameters: $\mu = np$ and $\sigma = \sqrt{np(1-p)}$

• Central Limit Theorem: Sums of independent random variables approach normal distribution

• CLT Sample Size: Works best with n ≄ 30, improves as n increases

• Sum Properties: Mean of sum = sum of means; Variance of sum = sum of variances (independent variables)

• Standardization Process: Convert any normal distribution to standard normal using z-scores, then use standard normal tables

Practice Quiz

5 questions to test your understanding