Cochran's theorem

From HandWiki

In statistics, Cochran's theorem, devised by William G. Cochran,[1] is a theorem used to justify results relating to the probability distributions of statistics that are used in the analysis of variance.[2]

Statement

Let U1, ..., UN be i.i.d. standard normally distributed random variables, and U=[U1,...,UN]T. Let B(1),B(2),,B(k)be symmetric matrices. Define ri to be the rank of B(i). Define Qi=UTB(i)U, so that the Qi are quadratic forms. Further assume iQi=UTU.

Cochran's theorem states that the following are equivalent:

Often it's stated as iAi=A, where A is idempotent, and iri=N is replaced by iri=rank(A). But after an orthogonal transform, A=diag(IM,0), and so we reduce to the above theorem.

Proof

Claim: Let X be a standard Gaussian in n, then for any symmetric matrices Q,Q, if XTQX and XTQX have the same distribution, then Q,Q have the same eigenvalues (up to multiplicity).

Claim: I=iBi.

Lemma: If iMi=I, all Mi symmetric, and have eigenvalues 0, 1, then they are simultaneously diagonalizable.

Now we prove the original theorem. We prove that the three cases are equivalent by proving that each case implies the next one in a cycle (1231).


Examples

Sample mean and sample variance

If X1, ..., Xn are independent normally distributed random variables with mean μ and standard deviation σ then

Ui=Xiμσ

is standard normal for each i. Note that the total Q is equal to sum of squared Us as shown here:

iQi=jikUjBjk(i)Uk=jkUjUkiBjk(i)=jkUjUkδjk=jUj2

which stems from the original assumption that B1+B2=I. So instead we will calculate this quantity and later separate it into Qi's. It is possible to write

i=1nUi2=i=1n(XiXσ)2+n(Xμσ)2

(here X is the sample mean). To see this identity, multiply throughout by σ2 and note that

(Xiμ)2=(XiX+Xμ)2

and expand to give

(Xiμ)2=(XiX)2+(Xμ)2+2(XiX)(Xμ).

The third term is zero because it is equal to a constant times

(XXi)=0,

and the second term has just n identical terms added together. Thus

(Xiμ)2=(XiX)2+n(Xμ)2,

and hence

(Xiμσ)2=(XiXσ)2+n(Xμσ)2=i(Ui1njUj)2Q1+1n(jUj)2Q2=Q1+Q2.

Now B(2)=Jnn with Jn the matrix of ones which has rank 1. In turn B(1)=InJnn given that In=B(1)+B(2). This expression can be also obtained by expanding Q1 in matrix notation. It can be shown that the rank of B(1) is n1 as the addition of all its rows is equal to zero. Thus the conditions for Cochran's theorem are met.

Cochran's theorem then states that Q1 and Q2 are independent, with chi-squared distributions with n − 1 and 1 degree of freedom respectively. This shows that the sample mean and sample variance are independent. This can also be shown by Basu's theorem, and in fact this property characterizes the normal distribution – for no other distribution are the sample mean and sample variance independent.[4]

Distributions

The result for the distributions is written symbolically as

(XiX)2σ2χn12.
n(Xμ)2σ2χ12,

Both these random variables are proportional to the true but unknown variance σ2. Thus their ratio does not depend on σ2 and, because they are statistically independent. The distribution of their ratio is given by

n(Xμ)21n1(XiX)2χ121n1χn12F1,n1

where F1,n − 1 is the F-distribution with 1 and n − 1 degrees of freedom (see also Student's t-distribution). The final step here is effectively the definition of a random variable having the F-distribution.

Estimation of variance

To estimate the variance σ2, one estimator that is sometimes used is the maximum likelihood estimator of the variance of a normal distribution

σ^2=1n(XiX)2.

Cochran's theorem shows that

nσ^2σ2χn12

and the properties of the chi-squared distribution show that

E(nσ^2σ2)=E(χn12)nσ2E(σ^2)=(n1)E(σ^2)=σ2(n1)n

Alternative formulation

The following version is often seen when considering linear regression.[5] Suppose that YNn(0,σ2In) is a standard multivariate normal random vector (here In denotes the n-by-n identity matrix), and if A1,,Ak are all n-by-n symmetric matrices with i=1kAi=In. Then, on defining ri=Rank(Ai), any one of the following conditions implies the other two:

  • i=1kri=n,
  • YTAiYσ2χri2 (thus the Ai are positive semidefinite)
  • YTAiY is independent of YTAjY for ij.

See also


References

  1. 1.0 1.1 Cochran, W. G. (April 1934). "The distribution of quadratic forms in a normal system, with applications to the analysis of covariance". Mathematical Proceedings of the Cambridge Philosophical Society 30 (2): 178–191. doi:10.1017/S0305004100016595. 
  2. Bapat, R. B. (2000). Linear Algebra and Linear Models (Second ed.). Springer. ISBN 978-0-387-98871-9. 
  3. "Cochran's theorem" (in en), A Dictionary of Statistics (Oxford University Press), 2008-01-01, doi:10.1093/acref/9780199541454.001.0001/acref-9780199541454-e-294, ISBN 978-0-19-954145-4, https://www.oxfordreference.com/view/10.1093/acref/9780199541454.001.0001/acref-9780199541454-e-294, retrieved 2022-05-18 
  4. "The Distribution of "Student's" Ratio for Non-Normal Samples". Supplement to the Journal of the Royal Statistical Society 3 (2): 178–184. 1936. doi:10.2307/2983669. 
  5. "Cochran's Theorem (A quick tutorial)". http://yangfeng.hosting.nyu.edu//slides/cochran's-theorem.pdf.