Notes on Principal Component Analysis
Notes on Principal Component Analysis
Notes on Principal Component Analysis
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
symmetry,<br />
or is an equicorrelati<strong>on</strong> matrix,<br />
⎛<br />
1 · · · ρ<br />
Σ p×p = σ 2 ⎜<br />
. .<br />
⎝<br />
ρ · · · 1<br />
P =<br />
⎛<br />
⎜<br />
⎝<br />
1 · · · ρ<br />
. .<br />
ρ · · · 1<br />
then the p − 1 smallest roots are all equal. For example, for the<br />
equicorrelati<strong>on</strong> matrix, λ 1 = 1 + (p − 1)ρ, and the remaining p −<br />
1 roots are all equal to 1 − ρ. The p × 1 eigenvector for λ 1 is<br />
[ √ 1<br />
p<br />
, . . . , √ 1<br />
p<br />
] ′ , and defines an average. Generally, for any variancecovariance<br />
matrix with all entries greater than zero (or just n<strong>on</strong>negative),<br />
the entries in the first eigenvector must all be greater than<br />
zero (or n<strong>on</strong>-negative). This is known as the Perr<strong>on</strong>-Frobenius theorem.<br />
Although we will not give these tests explicitly here (they can be<br />
found in Johns<strong>on</strong> and Wichern’s (2007) multivariate text), they are<br />
inference methods to test the null hypothesis of an equicorrelati<strong>on</strong><br />
matrix (i.e., the last p − 1 eigenvalues are equal); that the variancecovariance<br />
matrix is diag<strong>on</strong>al or the correlati<strong>on</strong> matrix is the identity<br />
(i.e., all eigenvalues are equal); or a sphericity test of independence<br />
that all eigenvalues are equal and Σ is σ 2 times the identity matrix.<br />
⎞<br />
⎟<br />
⎠<br />
⎞<br />
⎟<br />
⎠<br />
,<br />
,<br />
9