Steps to Carry Out The Experiment
- Identify the parameter of interest.
- Determine the null value and state the null hypothesis.
- State the appropriate alternative hypothesis.
- Give the formula of the computed value of the test statistic. Use the null value, but nothing from the sample!
- State the rejection region for your choice of \(\alpha\)
- Compute sample quantities, and calculate the value.
- Reject or not \(H_{0}\)
Very important: 2 and 3 above should be done prior to examining any data.
Case I: A Normal Population With Known \(\sigma\)
Rarely do we know \(\sigma\). But assume we do.
The Null Hypothesis: \(H_{0}:\mu=\mu_{0}\)
The test statistic: \(Z=\frac{\bar{X}-\mu_{0}}{\sigma/\sqrt{n}}\)
- If \(H_{a}:\mu>\mu_{0}\), the criterion is \(z\ge z_{\alpha}\) (you pick \(\alpha\)).
- If \(H_{a}:\mu<\mu_{0}\), the criterion is \(z\le-z_{\alpha}\)
- If \(H_{a}:\mu\ne\mu_{0}\), the criterion is \(z\ge z_{\alpha/2}\) or \(z\le-z_{\alpha/2}\)
\(\beta\) and Sample Size Determination
We usually don’t have a sample formula for \(\beta\), but we do for this case. Because we can do that, we can fix \(\alpha\) and \(\beta\) and calculate the needed \(n\).
For \(H_{a}:\mu>\mu_{0}\) and the actual \(\mu=\mu'\ne\mu_{0}\):
For \(H_{a}:\mu<\mu_{0}\) and \(\mu=\mu'\):
For \(H_{a}:\mu\ne\mu_{0}\) and \(\mu=\mu'\):
Given \(\alpha\) and \(\beta(\mu')=\beta\), \(n\) for one-tailed is:
And for two-tailed:
\(-z_{\beta}=z\) critical value that captures lower-tail \(\beta\):
for one-tailed.
Solve this equation for the desired \(n\).
For a fixed value \(\mu',\beta(\mu')\rightarrow0\) as \(n\rightarrow\infty\) for both one and two tailed for a normal population with known \(\sigma\).
Case II: Large Sample Tests
Let \(n>40\), and assume you don’t know the distribution or \(\sigma\). But we do know \(s\) is close to \(\sigma\).
So \(Z=\frac{\bar{X}-\mu}{s/\sqrt{n}}\) is approximately normal.
Use the rules for Case I.
Case III: A Normal Population Distribution
Whatr if \(n\) is small? We still go ahead with assuming a normal approximation (which may be a bad assumption). The main difference is that we now use the t-distribution.
\(H_{0}:\mu=\mu_{0}\)
- \(H_{a}:\mu>\mu_{0}\implies t\ge t_{\alpha,n-1}\)
- \(H_{a}:\mu<\mu_{0}\implies t\le-t_{\alpha,n-1}\)
- \(H_{a}:\mu\ne\mu_{0}\implies t\ge t_{\alpha/2,n-1}\) or \(t\le-t_{\alpha/2,n-1}\)
\(\beta\) and Sample Size Distribution
For a t-distribution, to determine \(n\) given \(\beta\), consult the relevant plots or tables.
Misc
If the population is normal and \(n\) is large, then \(S\) is approximately normal with \(E(S)\approx\sigma\) and \(V(S)\approx\frac{\sigma^{2}}{2n}\). Then \(\bar{X}\) and \(S\) are independent random variables.