12.07.2015 Views

Jolliffe I. Principal Component Analysis (2ed., Springer, 2002)(518s)

Jolliffe I. Principal Component Analysis (2ed., Springer, 2002)(518s)

Jolliffe I. Principal Component Analysis (2ed., Springer, 2002)(518s)

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

1IntroductionThe central idea of principal component analysis (PCA) is to reduce thedimensionality of a data set consisting of a large number of interrelatedvariables, while retaining as much as possible of the variation present inthe data set. This is achieved by transforming to a new set of variables,the principal components (PCs), which are uncorrelated, and which areordered so that the first few retain most of the variation present in all ofthe original variables.The present introductory chapter is in two parts. In the first, PCA isdefined, and what has become the standard derivation of PCs, in terms ofeigenvectors of a covariance matrix, is presented. The second part gives abrief historical review of the development of PCA.1.1 Definition and Derivation of<strong>Principal</strong> <strong>Component</strong>sSuppose that x is a vector of p random variables, and that the variancesof the p random variables and the structure of the covariances or correlationsbetween the p variables are of interest. Unless p is small, or thestructure is very simple, it will often not be very helpful to simply lookat the p variances and all of the 1 2p(p − 1) correlations or covariances. Analternative approach is to look for a few (≪ p) derived variables that preservemost of the information given by these variances and correlations orcovariances.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!