Chapter 8: Probability and Sampling Error
2026-02-11
This presentation is based on the following books. The references are coming from these books unless otherwise specified.
Main sources:
ClassShare App
You may be asked in class to go to the ClassShare App to answer questions.
SPSS Tutorial
By the end of this chapter, you should be able to:
| Symbol | Name | Pronunciation | Definition |
|---|---|---|---|
| \(\mu\) | Population mean | “myoo” | True average of the population |
| \(\bar{x}\) | Sample mean | “x bar” | Average of the sample |
| \(\sigma\) | Population standard deviation | “\(\sigma\)” | Population variability |
| \(s\) | Sample standard deviation | “s” | Sample variability |
| \(n\) | Sample size | “n” | Number of observations in a sample |
| \(SE\) | Standard error | “standard error” | Standard deviation of the sampling distribution |
| \(CI\) | Confidence interval | “C.I.” | Range of plausible values for a parameter |
| \(P(A)\) | Probability of event A | “probability of A” | Likelihood of event A occurring |
| \(z\) | Z-score | “zee” | Number of standard deviations from the mean |
In the real world, you don’t see the whole population. You only see one sample.
Scenario: You measure vertical jump height in 30 students.
Critical Questions:
The Challenge
We have this one number (\(\bar{x} = 51.2\)), but we know that sampling variability exists. How do we go from this single snapshot to making a confident statement about the entire population?
Statistical inference is the process of drawing conclusions about a population based on information from a sample[1].
Probability quantifies the likelihood that an event will occur, expressed as a number between 0 and 1[1].
Key concepts:
A is impossibleA is certainA has a 50-50 chanceTypes of probability:
Movement Science example:
If vertical jump heights are normally distributed with \(\mu = 50\) cm and \(\sigma = 8\) cm, what is the probability of jumping higher than 58 cm?
\[z = \frac{58 - 50}{8} = 1.0\]
\(P(X > 58) = 1 - 0.8413 = 0.1587\) → 15.87%
0.8413 is the cumulative probability up to 1 standard deviation above the mean.
Real-World Context
In a study of 200 college athletes, approximately 32 would be expected to jump higher than 58 cm (15.87% × 200 ≈ 32).
Answer: There is a 5% chance (or 1 in 20) that event A will occur. In statistics, this is the conventional threshold for “rare enough to be noteworthy” — the significance level α = 0.05.
Probability formalizes statements about uncertainty and enables evidence-based predictions.
Key Applications:
Why it matters
Research inherently involves uncertainty (sampling, measurement error, uncontrolled factors). Probability gives us the language to reason about this uncertainty.
Sampling error is the natural,unavoidable difference between a sample statistic (e.g., \(\bar{x}\)) and the true population parameter (e.g., \(\mu\))[1,2].
Key points:
Formula:
\[\text{Sampling Error} = \bar{x} - \mu\]
Important
Sampling error is not an error in the colloquial sense — it is the expected variability that arises from using a sample to estimate a population parameter.
Use this sequence when interpreting sample-based research findings:
The sampling distribution (distribution of sample means) is the distribution of a statistic (like the sample mean) across all possible samples of the same size from a population[1].
The bridge between descriptive statistics (describing our sample) and inferential statistics (making claims about populations) is built on sampling distributions[1].
Key properties:
Think of it this way
If you could take an infinite number of samples (each of size \(n\)) from the same population and calculate the mean of each, the distribution of those means would form the sampling distribution.
Answer: No! Sampling error is the natural, expected variability between a sample statistic and the population parameter. It occurs even with perfect random sampling because each sample contains different individuals.
The standard error (SE) is the standard deviation of the sampling distribution — it tells us how much sample means vary from sample to sample[1,2].
\[ SE = \frac{\sigma}{\sqrt{n}} \]
Key insights:
In real research, we rarely know the true population standard deviation (\(\sigma\)). Instead, we estimate it using the sample standard deviation (\(s\)).
\[ SE_{\bar{x}} \approx \frac{s}{\sqrt{n}} \]
The Power of Sample Size
The table below shows how the SE shrinks as sample size grows (assuming \(s = 10\)).
| Sample Size (\(n\)) | Standard Error (\(SE\)) |
|---|---|
| 10 | 3.16 |
| 30 | 1.83 |
| 100 | 1.00 |
| 400 | 0.50 |
The “Square Root” Rule: Because \(n\) is in the denominator under a square root, to cut the error in half, you must quadruple (\(4\times\)) the sample size.
Important
Diminishing Returns: Notice that the curve flattens out. Increasing \(n\) from 10 to 30 gives a huge gain in precision. Increasing \(n\) from 100 to 120 gives very little. This is crucial for cost-effective study design.
The Central Limit Theorem (CLT) is one of the most important theorems in statistics[1,2].
CLT: Regardless of the shape of the population distribution, the sampling distribution of the mean (sampling distribution of \(\bar{x}\)) approaches a normal distribution as the sample size increases.
Warning
We are not saying that sample means (your data) are normal. We are saying that the distribution of sample means (sampling distribution) is normal.
Conditions:
Implications:
Important
The CLT explains why statistics works: Even when we don’t know the shape of the population, we can make valid inferences about the mean as long as our sample is large enough.
Answer: No! That’s the beauty of the CLT — the sampling distribution of the mean approaches normality regardless of the population’s shape, as long as the sample size is sufficiently large.
A confidence interval (CI) is a range of plausible values for a population parameter, constructed from sample data[1,2].
Formula for 95% CI:
\[ CI_{95\%} = \bar{x} \pm 1.96 \times SE \]
Components:
Example: \(\bar{x} = 53\) cm, \(SE = 1.1\) cm
\[CI_{95\%} = 53 \pm 1.96(1.1) = [50.84, 55.16]\]
Important
Interpretation: “We are 95% confident that the true population mean falls between 50.84 and 55.16 cm.” This means that if we repeated this process many times, about 95% of our intervals would capture the true μ.
Answer: About 95 out of 100 intervals would be expected to contain the true population mean μ. Approximately 5 intervals would miss it — this is the inherent 5% error rate of 95% confidence intervals.
The width of a confidence interval is influenced by three factors[1]:
1. Sample size (\(n\)):
2. Variability (\(\sigma\) or \(s\)):
3. Confidence level:
| Confidence Level | Z-value | Interval Width |
|---|---|---|
| 90% | 1.645 | Narrowest |
| 95% | 1.960 | Moderate |
| 99% | 2.576 | Widest |
Trade-off
Higher confidence requires wider intervals (less precision). The 95% level is a convention that balances confidence and precision.
Point estimate: A single value used to estimate a population parameter
Interval estimate: A range of plausible values
APA Recommendation
The American Psychological Association (APA) recommends reporting confidence intervals alongside or instead of p-values[3].
Why? Confidence intervals provide more information:
Important
These concepts form the foundation for hypothesis testing (Chapter 10), where we’ll use probability to make formal decisions about whether observed differences are real or due to chance.