97.1.1 Statistical Hypothesis: Testing the Variance - Population
The population distribution of the random variable \(X\) is written as \(X \sim \text{N} \left( \mu, \sigma^2 \right)\) where \(\mu\) and \(\sigma^2\) represent the mean and variance of the normal distribution. In this representation it is assumed that \(\sigma^2\) is unknown. The parameter \(\mu\) can be either known or unknown.
97.1.2 Statistical Hypothesis: Testing the Variance - Sample
The statistic for the sample mean is \(\bar{x} = \frac{1}{n} \sum_{i=1}^n x_i\) where \(n\) is the number of observations in the sample. The sample statistic for the variance can be written in terms of \(\mu\) (if this population parameter is known) or in terms of the sample mean \(\bar{x}\):
The distribution of the sample variance can be written in terms of \(\mu\) (if this population parameter is known) or in terms of the sample mean \(\bar{x}\):
The sum of two independent \(\chi^2\)-distributed variates, with degrees of freedom \(n_1\) and \(n_2\) respectively, is also \(\chi^2\)-distributed with degrees of freedom equal to \(n_1 + n_2\). In general, the difference of two independent \(\chi^2\) variates is not \(\chi^2\)-distributed.
97.1.4.3 Property 2
The expected value of a \(\chi^2\)-distributed variate is equal to the number of degrees of freedom:
\[
\text{E} \left( \chi_n^2 \right) = n
\]
The variance of a \(\chi^2\)-distributed variate is equal to two times the number of degrees of freedom:
\[
\text{V} \left( \chi_n^2 \right) = 2n
\]
97.1.5 Approximation of the Chi-squared distribution
97.1.5.1 Rule of thumb
For large samples, the distribution of
\[
\sqrt{2 \chi_n^2} - \sqrt{2 n - 1}
\]
can be approximated by the standard normal distribution N\((0,1)\).
97.1.5.2 Example
Let \(n = 30\) and find the value \(c\) for which P\(\left( \chi_n^2 \geq c \right) = 0.05\).
Since \(k = 1.645\) it follows that the approximation results in \(c = 43.49\). According to the \(\chi^2\)-table, the correct value for the critical value is 43.773 (Appendix G). The approximation converges towards the correct value as \(n \rightarrow +\infty\).
From a random sample, with sample size \(n\) and drawn from a population following a normal distribution and given mean and standard deviation, the sample variance can be estimated as described in the following cases.
An interesting consequence of the previous case is that the statistic \(\frac{ns^2}{\sigma^2} = \frac{n \frac{1}{n}\sum_{i=1}^{n}\left( x_i - \mu \right)^2}{\sigma^2}\) is also \(\chi^2\)-distributed but with \(n\) degrees of freedom instead of \(n-1\). The loss of one degree of freedom in the first case is due to the substitution of the unknown population parameter \(\mu\) by the sample mean \(\bar{x}\).
97.1.7 Summary
Table 97.2: Estimation of Variance -- Test Statistics & Distributions
Hence P\(\left( \chi_7^2 \geq 12.44 \right) = 0.0869\). Note: the exact p-value cannot be determined based on Appendix G (it is only possible to use an approximate interpolation). With the use of statistical software, however, it is possible to obtain the exact p-value.
97.2.2.3 Conclusion
Since the probability \(0.0869\) is larger than \(\alpha = 0.05\) there is no reason to reject the Null Hypothesis.
The two-sided 90% acceptance region for the Sample Variance \(s^2\) (under H\(_0\)) is \([0.11146, 0.72345]\). In other words, P\(\left( 0.11146 \leq s^2 \leq 0.72345 \right) = 0.90\).
The two-sided 90% Confidence Interval for the Population Variance \(\sigma^2\) is \([0.31847, 2.06704]\). In other words, P\(\left( 0.31847 \leq \sigma^2 \leq 2.06704 \right) = 0.90\).