6. Statistical Inference

Tests Means

One-sample and two-sample t-tests, assumptions, pooled vs unpooled variance, and interpretation of results.

Tests of Means

Hey students! šŸ‘‹ Welcome to one of the most important topics in GCSE statistics - testing means! In this lesson, you'll discover how statisticians determine whether differences between groups are real or just due to random chance. We'll explore one-sample and two-sample t-tests, understand their assumptions, learn about pooled versus unpooled variance, and master how to interpret results. By the end, you'll have the statistical superpowers to analyze data like a pro! šŸš€

Understanding T-Tests: Your Statistical Detective Tool

Think of t-tests as your statistical detective tool šŸ”. Just like a detective needs evidence to prove a case, statisticians need t-tests to prove whether observed differences in data are meaningful or just coincidental.

A t-test is a statistical method used to compare means when we're working with small sample sizes (typically less than 30) or when we don't know the population standard deviation. The "t" comes from the t-distribution, which was discovered by William Gosset in 1908 while working for the Guinness brewery - yes, the beer company! šŸŗ

The t-distribution looks similar to the normal distribution but has thicker tails, which accounts for the extra uncertainty we have with smaller samples. As sample sizes increase, the t-distribution approaches the normal distribution.

Real-world example: Imagine you're testing whether a new energy drink actually improves athletic performance. You can't test the entire population, so you take a sample of 20 athletes. A t-test helps you determine if the performance improvement you observed is statistically significant or could have happened by chance.

One-Sample T-Tests: Comparing Against a Known Standard

A one-sample t-test compares the mean of a single group to a known or hypothetical value. This is like asking: "Is my sample different from what we expect?"

The formula for a one-sample t-test is:

$$t = \frac{\bar{x} - \mu_0}{s/\sqrt{n}}$$

Where:

  • $\bar{x}$ is the sample mean
  • $\mu_0$ is the hypothetical population mean
  • $s$ is the sample standard deviation
  • $n$ is the sample size

Real-world example: A chocolate manufacturer claims their bars weigh 100g on average. You randomly select 15 bars and find they weigh an average of 98.5g with a standard deviation of 2.1g. Is this significantly different from the claimed 100g?

Let's calculate: $t = \frac{98.5 - 100}{2.1/\sqrt{15}} = \frac{-1.5}{0.542} = -2.77$

With 14 degrees of freedom (n-1), this t-value suggests the bars are significantly lighter than claimed! šŸ“Š

The degrees of freedom for a one-sample t-test is always n-1. This accounts for the fact that we're using the sample standard deviation to estimate the population standard deviation.

Two-Sample T-Tests: Comparing Two Groups

A two-sample t-test compares the means of two independent groups. This answers questions like: "Are these two groups really different from each other?"

There are two main types of two-sample t-tests:

Independent Samples T-Test

Used when comparing two completely separate groups, like comparing test scores between Class A and Class B.

Paired Samples T-Test

Used when the same subjects are measured twice, like comparing before-and-after weights in a diet study.

Real-world example: A school wants to know if boys and girls perform differently on math tests. They randomly select 25 boys (mean = 78, SD = 12) and 23 girls (mean = 82, SD = 10). The two-sample t-test helps determine if this 4-point difference is statistically significant.

Pooled vs Unpooled Variance: Making the Right Choice

Here's where things get interesting, students! When conducting two-sample t-tests, you need to decide whether to use pooled or unpooled variance.

Pooled Variance

Pooled variance assumes both groups have equal population variances. It combines (pools) the variance estimates from both samples to create a single, more reliable estimate.

The pooled variance formula is:

$$s_p^2 = \frac{(n_1-1)s_1^2 + (n_2-1)s_2^2}{n_1 + n_2 - 2}$$

Use pooled variance when:

  • Sample sizes are approximately equal
  • The variances appear similar
  • You have reason to believe the populations have equal variances

Unpooled Variance (Welch's T-Test)

Unpooled variance doesn't assume equal variances and calculates separate variance estimates for each group.

Use unpooled variance when:

  • Sample sizes are very different
  • The variances appear quite different
  • You're unsure about the equal variance assumption

Rule of thumb: If one sample variance is more than twice the other, consider using unpooled variance! šŸ“

Key Assumptions: The Foundation of Valid Results

T-tests aren't magic - they only work properly when certain assumptions are met:

1. Normality

The data should be approximately normally distributed. With larger samples (n > 30), this becomes less critical due to the Central Limit Theorem.

How to check: Create histograms or Q-Q plots of your data. Look for roughly bell-shaped distributions.

2. Independence

Each observation should be independent of others. This means one person's score shouldn't influence another's.

Real-world violation: Testing students sitting next to each other who might copy answers.

3. Equal Variances (for pooled tests)

When using pooled variance, assume both groups have similar variability.

How to check: Compare sample standard deviations. If one is more than twice the other, consider unpooled methods.

4. Random Sampling

Your sample should be randomly selected from the population of interest.

When these assumptions are violated, your results might be unreliable. Don't worry though - there are alternative tests (like Mann-Whitney U) for when assumptions aren't met! šŸ›”ļø

Interpreting Results: Making Sense of Your Analysis

Understanding your t-test results is crucial, students! Here's what to look for:

P-Values

The p-value tells you the probability of getting your observed results (or more extreme) if there's actually no real difference.

  • p < 0.05: Statistically significant (strong evidence against null hypothesis)
  • p ≄ 0.05: Not statistically significant (insufficient evidence)

Effect Size

Statistical significance doesn't always mean practical importance. Cohen's d measures effect size:

$$d = \frac{\bar{x_1} - \bar{x_2}}{s_{pooled}}$$

$- Small effect: d = 0.2$

$- Medium effect: d = 0.5 $

$- Large effect: d = 0.8$

Real-world example: A new teaching method improves test scores by 2 points (p < 0.01, d = 0.1). While statistically significant, the tiny effect size suggests it's not practically meaningful.

Confidence Intervals

A 95% confidence interval shows the range where the true difference likely lies. If this interval doesn't include zero, you have significant results.

Conclusion

Congratulations, students! šŸŽ‰ You've mastered the fundamentals of testing means through t-tests. We've explored how one-sample t-tests compare a group to a known value, while two-sample t-tests compare two groups. You've learned the critical decision between pooled and unpooled variance methods, understood the essential assumptions that make t-tests valid, and discovered how to interpret results meaningfully. Remember, statistical significance (p-values) tells you if an effect exists, but effect size tells you if it matters practically. These tools will serve you well in analyzing real-world data and making evidence-based decisions throughout your statistical journey!

Study Notes

• One-sample t-test: Compares sample mean to known value using $t = \frac{\bar{x} - \mu_0}{s/\sqrt{n}}$

• Two-sample t-test: Compares means of two independent groups

• Degrees of freedom: One-sample = n-1, Two-sample pooled = n₁ + nā‚‚ - 2

• Pooled variance: Use when sample sizes equal and variances similar

• Unpooled variance: Use when sample sizes very different or variances differ significantly

• Key assumptions: Normality, independence, equal variances (for pooled), random sampling

• P-value < 0.05: Statistically significant result

• Cohen's d effect sizes: Small (0.2), Medium (0.5), Large (0.8)

• Confidence intervals: If 95% CI doesn't include zero, result is significant

• T-distribution: Similar to normal but with thicker tails for small samples

• Rule of thumb: Use unpooled if one variance > 2Ɨ the other variance

Practice Quiz

5 questions to test your understanding

Tests Means — GCSE Statistics | A-Warded