Understanding Random Variation¶
When we make measurements, we often observe that repeated measurements of the same quantity show random variations. This is a fundamental aspect of experimental science that we must understand and account for.
Consider measuring the radioactivity of a sample. Even with perfect equipment, the number of counts in a fixed time interval will vary randomly due to the inherent stochastic nature of radioactive decay. Similarly, optical measurements might show fluctuations due to air currents causing refractive index variations or thermal effects causing mechanical instabilities in the apparatus.
The Gaussian Distribution¶
When we make many measurements of the same quantity, the results often follow a bell-shaped curve known as the Gaussian or normal distribution. This distribution is fundamental to understanding measurement uncertainty.
The mathematical form of the Gaussian distribution is:
where μ is the population mean and σ is the population standard deviation.
These percentages are crucial for understanding measurement uncertainty. The 68-95-99.7 rule (sometimes called the empirical rule) provides a quick way to assess the likelihood that a measurement falls within certain bounds of the true value.
This distribution allows us to make meaningful statements about our measurements. For example, if we measure a length multiple times and find a mean of 10.5 cm with a standard deviation of 0.1 cm, we can say:
Sample Statistics and Population Parameters¶
When we make measurements, we’re typically working with a sample from a larger population of possible measurements. Understanding the relationship between sample statistics and population parameters is essential.
The sample mean is calculated as:
The sample standard deviation is:
where is the sample size and are individual measurements.
Note the in the denominator, known as Bessel’s correction, which provides an unbiased estimate of the population standard deviation.
The standard error of the mean tells us how precisely we’ve determined the population mean. As our sample size increases, this uncertainty decreases as .
Distinction Between Standard Deviation and Standard Error¶
This distinction is crucial and frequently misunderstood:
Standard Deviation (): Describes the variability of individual measurements around the sample mean. It tells us about the inherent scatter in our data.
Standard Error of the Mean (): Describes the uncertainty in our estimate of the population mean. It tells us how precisely we know the “true” value.
For example, if we measure the same quantity 25 times and get units:
- The standard deviation remains 2.0 units (describing individual measurement scatter)
- The standard error of the mean is units (our uncertainty in the mean)
Propagation of Statistical Uncertainty¶
When we calculate derived quantities from multiple measurements, we need to understand how the uncertainties combine. The propagation formulas depend on whether we’re dealing with estimated uncertainties or statistical uncertainties.
General Error Propagation Rules¶
For a function where and are independent variables with standard deviations and :
example - Unknown Directive
example - Unknown DirectiveIf we measure a rectangle's length as $(25.4 \pm 0.2)$ cm and width as $(18.6 \pm 0.2)$ cm, what is the uncertainty in the area?
**Solution**:
1. Calculate the area: $A = L \times W = 25.4 \text{cm} \times 18.6 \text{cm} = 472.44 \text{cm}^2$
2. Calculate relative uncertainties:
- $\frac{\sigma_L}{L} = \frac{0.2}{25.4} \approx 0.00787$
- $\frac{\sigma_W}{W} = \frac{0.2}{18.6} \approx 0.01075$
3. Combine relative uncertainties: Then use the multiplication rule:
$$\left(\frac{\sigma_A}{A}\right)^2 = (0.00787)^2 + (0.01075)^2 \approx 0.0000619 + 0.0001156 = 0.0001775$$
$$\frac{\sigma_A}{A} = \sqrt{0.0001775} \approx 0.0133$$
$$\sigma_A = 0.0133 \times 472.44 \approx
$$\sigma_A = \sqrt{0.0001775} \times 472.44 \text{cm}^2 \approx 6.3 \ \text{cm}^2$$
4. Final result: $A = (472 \pm 6) \ \text{cm}^2$Statistical vs. Estimated Uncertainties¶
When combining different types of uncertainties (e.g., some estimated, some statistical), we need to ensure compatibility. If one uncertainty represents a 68% confidence interval (1 standard deviation) and another represents outer limits (~100% confidence), they cannot be directly combined using the standard propagation formulas.
Central Limit Theorem and Sampling¶
The Central Limit Theorem explains why the Gaussian distribution is so prevalent in measurement science. It states that the distribution of sample means approaches a normal distribution as the sample size increases, regardless of the shape of the original population distribution.
This theorem justifies our use of Gaussian statistics even when individual measurements might not follow a perfect Gaussian distribution.
Identifying and Handling Outliers¶
Sometimes our measurements include values that seem unusually different from the others. These outliers require careful consideration and systematic analysis.
Chauvenet’s Criterion¶
Chauvenet’s criterion provides a statistical method for identifying potential outliers. The criterion states that a measurement should be rejected if the probability of obtaining a deviation as large or larger is less than , where is the total number of measurements.
Procedure for Chauvenet’s Criterion:
- Calculate the sample mean () and standard deviation ()
- For each measurement , calculate the deviation:
- Express this as a number of standard deviations:
- Find the probability that a measurement would deviate by or more standard deviations (using Gaussian tables)
- If this probability is less than , the measurement is a candidate for rejection
Example: For measurements, reject if probability < 0.05 (about ) For measurements, reject if probability < 0.025 (about )
Systematic Approach to Outliers¶
When handling potential outliers, follow this systematic approach:
- Verify the measurement was recorded correctly
- Check for obvious experimental problems (equipment malfunction, environmental disturbance)
- Apply statistical assessment (such as Chauvenet’s criterion)
- Document thoroughly the reasoning for any rejected measurements
- Never reject data simply because it doesn’t fit expectations
The probability guidelines from the Gaussian distribution help us make these decisions:
- Outside limits: 5% probability (might be legitimate outlier)
- Outside limits: 0.3% probability (likely candidate for rejection)
- Outside limits: 0.006% probability (very likely experimental error)
Confidence Intervals and Uncertainty Statements¶
Understanding how to make proper uncertainty statements is crucial for communicating experimental results.
Confidence Intervals¶
A confidence interval provides a range of values that likely contains the true population parameter. For a 95% confidence interval of the mean:
This means we’re 95% confident the true population mean lies within this range.
Proper Uncertainty Statements¶
When reporting results:
- 68% confidence: (1 standard error)
- 95% confidence: (≈ 2 standard errors)
- 99.7% confidence: (3 standard errors)
Sample Size Effects¶
The size of our sample dramatically affects the reliability of our statistical estimates.
Effect on Standard Error¶
The standard error of the mean decreases as :
- To halve the uncertainty in the mean, need 4 times as many measurements
- To reduce uncertainty by factor of 10, need 100 times as many measurements
Reliability of Standard Deviation Estimates¶
For small samples, our estimate of the population standard deviation is quite uncertain. The standard deviation of the standard deviation is approximately:
This means:
- With : our estimate has ~35% uncertainty
- With : our estimate has ~25% uncertainty
- With : our estimate has ~15% uncertainty
Combining Different Types of Uncertainty¶
In practice, we often need to combine uncertainties that have different statistical meanings (e.g., some estimated, some statistical).
Making Uncertainties Compatible¶
If combining a statistically-based uncertainty (68% confidence) with an estimated range uncertainty (~100% confidence), we need to make them compatible:
- Convert estimated range to ~68% confidence by multiplying by 0.67
- Or convert statistical uncertainty to ~100% range by multiplying by 1.5
Root Sum of Squares¶
For independent uncertainties of the same confidence level:
This assumes the uncertainties are:
- Independent (not correlated)
- Random (not systematic)
- Of the same confidence level
Distribution Shapes and Assumptions¶
While we often assume Gaussian distributions, real measurements may deviate from this ideal.
When Gaussian Assumptions Fail¶
- Small number statistics: For counting experiments with few counts, use Poisson statistics
- Skewed distributions: May occur with certain measurement processes
- Multiple peaks: Could indicate multiple measurement modes or systematic effects
Checking Gaussian Assumptions¶
Simple tests for Gaussian behavior:
- Plot histogram of residuals from the mean
- Check if ~68% fall within 1 standard deviation
- Look for systematic patterns in the data
Practical Measurement Strategy¶
A well-planned measurement strategy can minimize uncertainties and improve data quality.
Before starting measurements:
- Estimate expected uncertainty based on instrument resolution and known fluctuations
- Determine required sample size based on target precision
- Choose measurement sequence to minimize systematic effects
- Plan for outlier detection and handling procedures
Glossary¶
- Gaussian distribution
- A bell-shaped probability distribution that describes many random phenomena in nature.
- standard deviation
- A measure of the spread of a distribution, indicating how much variation exists from the mean.
- standard error
- The standard deviation of the sampling distribution of a statistic, most commonly the standard error of the mean.
- outlier
- A data point that appears to deviate markedly from other observations in a dataset.
- confidence interval
- A range of values that is likely to contain the true value of a parameter with a specified level of confidence.
- propagation of uncertainty
- The process of determining how uncertainties in individual measurements affect the uncertainty in a calculated result.
- Chauvenet’s criterion
- A statistical rule for identifying potential outliers based on the probability of observing such deviations.
- sample mean
- The arithmetic average of a set of sample values, used to estimate the population mean.
- population mean
- The true average value of a quantity if we could measure it an infinite number of times.
- central limit theorem
- A statistical theorem stating that sample means approach a normal distribution as sample size increases, regardless of the original population distribution.