Jolliffe I. Principal Component Analysis (2ed., Springer, 2002)(518s)
Jolliffe I. Principal Component Analysis (2ed., Springer, 2002)(518s) Jolliffe I. Principal Component Analysis (2ed., Springer, 2002)(518s)
6Choosing a Subset of PrincipalComponents or VariablesIn this chapter two separate, but related, topics are considered, both ofwhich are concerned with choosing a subset of variables. In the first section,the choice to be examined is how many PCs adequately account for thetotal variation in x. The major objective in many applications of PCA isto replace the p elements of x by a much smaller number m of PCs, whichnevertheless discard very little information. It is crucial to know how smallm can be taken without serious information loss. Various rules, many adhoc, have been proposed for determining a suitable value of m, and theseare discussed in Section 6.1. Examples of their use are given in Section 6.2.Using m PCs instead of p variables considerably reduces the dimensionalityof the problem when m ≪ p, but usually the values of all p variablesare still needed in order to calculate the PCs, as each PC is likely to bea function of all p variables. It might be preferable if, instead of using mPCs we could use m, or perhaps slightly more, of the original variables,to account for most of the variation in x. The question arises of how tocompare the information contained in a subset of variables with that inthe full data set. Different answers to this question lead to different criteriaand different algorithms for choosing the subset. In Section 6.3 we concentrateon methods that either use PCA to choose the variables or aim toreproduce the PCs in the full data set with a subset of variables, thoughother variable selection techniques are also mentioned briefly. Section 6.4gives two examples of the use of variable selection methods.All of the variable selection methods described in the present chapterare appropriate when the objective is to describe variation within x aswell as possible. Variable selection when x is a set of regressor variables
112 6. Choosing a Subset of Principal Components or Variablesin a regression analysis, or a set of predictor variables in a discriminantanalysis, is a different type of problem as criteria external to x must beconsidered. Variable selection in regression is the subject of Section 8.5. Therelated problem of choosing which PCs to include in a regression analysisor discriminant analysis is discussed in Sections 8.2, 9.1 respectively.6.1 How Many Principal Components?In this section we present a number of rules for deciding how many PCsshould be retained in order to account for most of the variation in x (orin the standardized variables x ∗ in the case of a correlation matrix-basedPCA).In some circumstances the last few, rather than the first few, PCs are ofinterest, as was discussed in Section 3.4 (see also Sections 3.7, 6.3, 8.4, 8.6and 10.1). In the present section, however, the traditional idea of tryingto reduce dimensionality by replacing the p variables by the first m PCs(m
- Page 92 and 93: 3.9. Models for Principal Component
- Page 94 and 95: 4Principal Components as a SmallNum
- Page 96 and 97: 4.1. Anatomical Measurements 65Tabl
- Page 98 and 99: 4.1. Anatomical Measurements 67spac
- Page 100 and 101: 4.2. The Elderly at Home 69Table 4.
- Page 102 and 103: 4.3. Spatial and Temporal Variation
- Page 104 and 105: 4.3. Spatial and Temporal Variation
- Page 106 and 107: 4.4. Properties of Chemical Compoun
- Page 108 and 109: 4.5. Stock Market Prices 77Table 4.
- Page 110 and 111: 5. Graphical Representation of Data
- Page 112 and 113: Anatomical Measurements5.1. Plottin
- Page 114 and 115: 5.1. Plotting Two or Three Principa
- Page 116 and 117: 5.2. Principal Coordinate Analysis
- Page 118 and 119: 5.2. Principal Coordinate Analysis
- Page 120 and 121: 5.2. Principal Coordinate Analysis
- Page 122 and 123: 5.3. Biplots 91columns, L is an (r
- Page 124 and 125: 5.3. Biplots 93ButandSubstituting i
- Page 126 and 127: 5.3. Biplots 95The vector gi ∗ co
- Page 128 and 129: 5.3. Biplots 97Figure 5.3. Biplot u
- Page 130 and 131: 5.3. Biplots 99Table 5.2. First two
- Page 132 and 133: 5.3. Biplots 101Figure 5.5. Biplot
- Page 134 and 135: 5.4. Correspondence Analysis 103of
- Page 136 and 137: 5.4. Correspondence Analysis 105Fig
- Page 138 and 139: 5.6. Displaying Intrinsically High-
- Page 140 and 141: 5.6. Displaying Intrinsically High-
- Page 144 and 145: 6.1. How Many Principal Components?
- Page 146 and 147: 6.1. How Many Principal Components?
- Page 148 and 149: 6.1. How Many Principal Components?
- Page 150 and 151: 6.1. How Many Principal Components?
- Page 152 and 153: 6.1. How Many Principal Components?
- Page 154 and 155: 6.1. How Many Principal Components?
- Page 156 and 157: 6.1. How Many Principal Components?
- Page 158 and 159: 6.1. How Many Principal Components?
- Page 160 and 161: 6.1. How Many Principal Components?
- Page 162 and 163: 6.1. How Many Principal Components?
- Page 164 and 165: 6.2. Choosing m, the Number of Comp
- Page 166 and 167: 6.2. Choosing m, the Number of Comp
- Page 168 and 169: 6.3. Selecting a Subset of Variable
- Page 170 and 171: 6.3. Selecting a Subset of Variable
- Page 172 and 173: 6.3. Selecting a Subset of Variable
- Page 174 and 175: 6.3. Selecting a Subset of Variable
- Page 176 and 177: 6.4. Examples Illustrating Variable
- Page 178 and 179: 6.4. Examples Illustrating Variable
- Page 180 and 181: 6.4. Examples Illustrating Variable
- Page 182 and 183: 7.1. Models for Factor Analysis 151
- Page 184 and 185: 7.2. Estimation of the Factor Model
- Page 186 and 187: 7.2. Estimation of the Factor Model
- Page 188 and 189: 7.2. Estimation of the Factor Model
- Page 190 and 191: 7.3. Comparisons Between Factor and
112 6. Choosing a Subset of <strong>Principal</strong> <strong>Component</strong>s or Variablesin a regression analysis, or a set of predictor variables in a discriminantanalysis, is a different type of problem as criteria external to x must beconsidered. Variable selection in regression is the subject of Section 8.5. Therelated problem of choosing which PCs to include in a regression analysisor discriminant analysis is discussed in Sections 8.2, 9.1 respectively.6.1 How Many <strong>Principal</strong> <strong>Component</strong>s?In this section we present a number of rules for deciding how many PCsshould be retained in order to account for most of the variation in x (orin the standardized variables x ∗ in the case of a correlation matrix-basedPCA).In some circumstances the last few, rather than the first few, PCs are ofinterest, as was discussed in Section 3.4 (see also Sections 3.7, 6.3, 8.4, 8.6and 10.1). In the present section, however, the traditional idea of tryingto reduce dimensionality by replacing the p variables by the first m PCs(m