The purpose of confidence intervals lies in estimating things for our population. The basic train of logic is that our estimate (the corresponding sample statistic) will not be exactly right; it will be off a bit due to one or both of the following:
Confidence intervals take into account the sampling variability – the fact that each sample is slightly different and gives us a different statistic to estimate the population parameter. CIs provide a range of estimates that are ‘plausible’ for the parameter value.
If the population is normal, even with small sample sizes, CIs will work well. Most of the time we won’t know this is the case, so we check sample size. Random samples help remove bias.
Note: When we know the std. dev. of the population (\(\sigma\)), then we can just use the Normal distribution for our CIs. Otherwise, we must use the t-distribution.
value of \(z^*\) determined by confidence level.
Common values of \(z^*\):
where \(t_{(1-\alpha/2, df=n-1)}\)
is the \(1-\alpha/2\) quantile for a
t-distribution with n-1 degrees of freedom (use qt()
to get
final values).
Common values of \(1-\alpha/2\):
If the population is normal, even with small sample sizes, CIs will work well. Most of the time we won’t know this is the case, so we check sample size. Random samples help remove bias. We will pretty much never work with a scenario where \(\sigma\) for both groups is known, so I won’t post the formulas for that case.
where \(t_{(1-\alpha/2, df)}\) is
the \(1-\alpha/2\) quantile for a
t-distribution with degrees of freedom \(df=min(n_1, n_2)-1\) (use qt()
to get final values).
Common values of \(1-\alpha/2\):