Cochran's theorem
In statistics, Cochran's theorem, devised by William G. Cochran,[1] is a theorem used in to justify results relating to the probability distributions of statistics that are used in the analysis of variance.[2]
Contents
Statement
|
This article's factual accuracy is disputed. Please see the relevant discussion on the talk page. (July 2009) |
Suppose U1, ..., Un are independent standard normally distributed random variables, and an identity of the form
can be written, where each Qi is a sum of squares of linear combinations of the Us. Further suppose that
where ri is the rank of Qi. Cochran's theorem states that the Qi are independent, and each Qi has a chi-square distribution with ri degrees of freedom.
Examples
Sample mean and sample variance
If X1, ..., Xn are independent normally distributed random variables with mean μ and standard deviation σ then
is standard normal for each i.
It is possible to write
(here, summation is from 1 to n, that is over the observations). To see this identity, multiply throughout by σ2 and note that
and expand to give
The third term is zero because it is equal to a constant times
and the second term is just n identical terms added together.
Combining the above results (and dividing by σ2), we have:
Now the rank of Q2 is just 1 (it is the square of just one linear combination of the standard normal variables). The rank of Q1 can be shown to be n − 1, and thus the conditions for Cochran's theorem are met.
Cochran's theorem then states that Q1 and Q2 are independent, with Chi-squared distribution with n − 1 and 1 degree of freedom respectively.
This shows that the sample mean and sample variance are independent; this can also be show by Basu's theorem, and this property characterizes the normal distribution – for no other distribution are the sample mean and sample variance independent.
Further,
Estimation of variance
To estimate the variance σ2, one estimator that is often used is
Cochran's theorem shows that
which shows that the expected value of is σ2(n − 1)/n.
Both these distributions are proportional to the true but unknown variance σ2; thus their ratio is independent of σ2 and because they are independent we have
where F1,n − 1 is the F-distribution with 1 and n − 1 degrees of freedom (see also Student's t-distribution).
Alternative formulation
The following version is often seen when considering linear regression.
Suppose Y˜Nn(0,σ2In) is a standard multivariate Gaussian random variable [here In denotes the n-by-n identity matrix], and if are all n-by-n symmetric matrices with
. Then, writing Rank(Ai) = ri, any one of the following conditions implies the other two:
See also
- Cramér's theorem, on decomposing normal distribution
- Infinite divisibility (probability)
References
- ^ Cochran, W. G. (April 1934). "The distribution of quadratic forms in a normal system, with applications to the analysis of covariance". Mathematical Proceedings of the Cambridge Philosophical Society 30 (2): 178-191. doi:10.1017/S0305004100016595.
- ^ Bapat, R. B. (2000). Linear Algebra and Linear Models (Second ed.). Springer. ISBN 9780387988719.
This entry is from Wikipedia, the leading user-contributed encyclopedia. It may not have been reviewed by professional editors (see full disclaimer)