Research Methodology - Dr. Krishan K. Pandey
Research Methodology - Dr. Krishan K. Pandey Research Methodology - Dr. Krishan K. Pandey
Multivariate Analysis Techniques 331 Correlation Matrix, R Variables X 1 X 2 X 3 …. X k X 1 r 11 r 12 r 13 …. r 1k X 2 r 21 r 22 r 23 …. r 3k Variables X 3 r 31 r 32 r 33 …. r 3k . . . . . . . . X k r k1 r k2 r k3 …. r kk The main diagonal spaces include unities since such elements are self-correlations. The correlation matrix happens to be a symmetrical matrix. (ii) Presuming the correlation matrix to be positive manifold (if this is not so, then reflections as mentioned in case of centroid method must be made), the first step is to obtain the sum of coefficients in each column, including the diagonal element. The vector of column sums is referred to as U a1 and when U a1 is normalized, we call it V a1 . This is done by squaring and summing the column sums in U a1 and then dividing each element in U a1 by the square root of the sum of squares (which may be termed as normalizing factor). Then elements in V a1 are accumulatively multiplied by the first row of R to obtain the first element in a new vector U a2 . For instance, in multiplying V a1 by the first row of R, the first element in V a1 would be multiplied by the r 11 value and this would be added to the product of the second element in V a1 multiplied by the r 12 value, which would be added to the product of third element in V a1 multiplied by the r 13 value, and so on for all the corresponding elements in V a1 and the first row of R. To obtain the second element of U a2 , the same process would be repeated i.e., the elements in V a1 are accumulatively multiplied by the 2nd row of R. The same process would be repeated for each row of R and the result would be a new vector U a2 . Then U a2 would be normalized to obtain V a2 . One would then compare V a1 and V a2 . If they are nearly identical, then convergence is said to have occurred (If convergence does not occur, one should go on using these trial vectors again and again till convergence occurs). Suppose the convergence occurs when we work out V a8 in which case V a7 will be taken as V a (the characteristic vector) which can be converted into loadings on the first principal component when we multiply the said vector (i.e., each element of V a ) by the square root of the number we obtain for normalizing U a8 . (iii) To obtain factor B, one seeks solutions for V b , and the actual factor loadings for second component factor, B. The same procedures are used as we had adopted for finding the first factor, except that one operates off the first residual matrix, R 1 rather than the original correlation matrix R (We operate on R 1 in just the same way as we did in case of centroid method stated earlier). (iv) This very procedure is repeated over and over again to obtain the successive PC factors (viz. C, D, etc.).
332 Research Methodology Other steps involved in factor analysis (a) Next the question is: How many principal components to retain in a particular study? Various criteria for this purpose have been suggested, but one often used is Kaiser’s criterion. According to this criterion only the principal components, having latent root greater than one, are considered as essential and should be retained. (b) The principal components so extracted and retained are then rotated from their beginning position to enhance the interpretability of the factors. (c) Communality, symbolized, h2 , is then worked out which shows how much of each variable is accounted for by the underlying factors taken together. A high communality figure means that not much of the variable is left over after whatever the factors represent is taken into consideration. It is worked out in respect of each variable as under: h2 of the ith variable = (ith factor loading of factor A) 2 + (ith factor loading of factor B) 2 + … Then follows the task of interpretation. The amount of variance explained (sum of squared loadings) by each PC factor is equal to the corresponding characteristic root. When these roots are divided by the number of variables, they show the characteristic roots as proportions of total variance explained. (d) The variables are then regressed against each factor loading and the resulting regression coefficients are used to generate what are known as factor scores which are then used in further analysis and can also be used as inputs in several other multivariate analyses. Illustration 3 Take the correlation matrix, R, for eight variables of illustration 1 of this chapter and then compute: (i) the first two principal component factors; (ii) the communality for each variable on the basis of said two component factors; (iii) the proportion of total variance as well as the proportion of common variance explained by each of the two component factors. Solution: Since the given correlation matrix is a positive manifold, we work out the first principal component factor (using trial vectors) as under: Table 13.3 Variables 1 2 3 4 5 6 7 8 1 1.000 .709 .204 .081 .626 .113 .155 .774 2 .709 1.000 .051 .089 .581 .098 .083 .652 3 .204 .051 1.000 .671 .123 .689 .582 .072 4 .081 .089 .671 1.000 .022 .798 .613 .111 Variables 5 .626 .581 .123 .022 1.000 .047 .201 .724 6 .113 .098 .689 .798 .047 1.000 .801 .120 ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ ○ Contd.
- Page 298 and 299: Analysis of Variance and Co-varianc
- Page 300 and 301: Testing of Hypotheses-II 283 12 Tes
- Page 302 and 303: Testing of Hypotheses-II 285 Illust
- Page 304 and 305: Testing of Hypotheses-II 287 Table
- Page 306 and 307: Testing of Hypotheses-II 289 fact,
- Page 308 and 309: Testing of Hypotheses-II 291 The te
- Page 310 and 311: Testing of Hypotheses-II 293 Pair B
- Page 312 and 313: Testing of Hypotheses-II 295 Table
- Page 314 and 315: Testing of Hypotheses-II 297 Illust
- Page 316 and 317: Testing of Hypotheses-II 299 Soluti
- Page 318 and 319: Testing of Hypotheses-II 301 r = nu
- Page 320 and 321: Testing of Hypotheses-II 303 where
- Page 322 and 323: Testing of Hypotheses-II 305 Table
- Page 324 and 325: Testing of Hypotheses-II 307 We now
- Page 326 and 327: Testing of Hypotheses-II 309 (ii) I
- Page 328 and 329: Testing of Hypotheses-II 311 all po
- Page 330 and 331: Testing of Hypotheses-II 313 CONCLU
- Page 332 and 333: Multivariate Analysis Techniques 31
- Page 334 and 335: Multivariate Analysis Techniques 31
- Page 336 and 337: Multivariate Analysis Techniques 31
- Page 338 and 339: Multivariate Analysis Techniques 32
- Page 340 and 341: Multivariate Analysis Techniques 32
- Page 342 and 343: Multivariate Analysis Techniques 32
- Page 344 and 345: Multivariate Analysis Techniques 32
- Page 346 and 347: Multivariate Analysis Techniques 32
- Page 350 and 351: Multivariate Analysis Techniques 33
- Page 352 and 353: Multivariate Analysis Techniques 33
- Page 354 and 355: Multivariate Analysis Techniques 33
- Page 356 and 357: Multivariate Analysis Techniques 33
- Page 358 and 359: Multivariate Analysis Techniques 34
- Page 360 and 361: Appendix Summary Chart: Showing the
- Page 362 and 363: Interpretation and Report Writing 3
- Page 364 and 365: Interpretation and Report Writing 3
- Page 366 and 367: Interpretation and Report Writing 3
- Page 368 and 369: Interpretation and Report Writing 3
- Page 370 and 371: Interpretation and Report Writing 3
- Page 372 and 373: Interpretation and Report Writing 3
- Page 374 and 375: Interpretation and Report Writing 3
- Page 376 and 377: Interpretation and Report Writing 3
- Page 378 and 379: The Computer: Its Role in Research
- Page 380 and 381: The Computer: Its Role in Research
- Page 382 and 383: The Computer: Its Role in Research
- Page 384 and 385: The Computer: Its Role in Research
- Page 386 and 387: The Computer: Its Role in Research
- Page 388 and 389: The Computer: Its Role in Research
- Page 390 and 391: The Computer: Its Role in Research
- Page 392 and 393: Appendix 375 Appendix (Selected Sta
- Page 394 and 395: Appendix 377 Table 2: Critical Valu
- Page 396 and 397: Appendix 379 v 2 Table 4(a): Critic
Multivariate Analysis Techniques 331<br />
Correlation Matrix, R<br />
Variables<br />
X 1 X 2 X 3 …. X k<br />
X 1 r 11 r 12 r 13 …. r 1k<br />
X 2 r 21 r 22 r 23 …. r 3k<br />
Variables X 3 r 31 r 32 r 33 …. r 3k<br />
. . . .<br />
. . . .<br />
X k r k1 r k2 r k3 …. r kk<br />
The main diagonal spaces include unities since such elements are self-correlations. The<br />
correlation matrix happens to be a symmetrical matrix.<br />
(ii) Presuming the correlation matrix to be positive manifold (if this is not so, then reflections as<br />
mentioned in case of centroid method must be made), the first step is to obtain the sum of<br />
coefficients in each column, including the diagonal element. The vector of column sums is<br />
referred to as U a1 and when U a1 is normalized, we call it V a1 . This is done by squaring and<br />
summing the column sums in U a1 and then dividing each element in U a1 by the square root<br />
of the sum of squares (which may be termed as normalizing factor). Then elements in V a1<br />
are accumulatively multiplied by the first row of R to obtain the first element in a new<br />
vector U a2 . For instance, in multiplying V a1 by the first row of R, the first element in V a1<br />
would be multiplied by the r 11 value and this would be added to the product of the second<br />
element in V a1 multiplied by the r 12 value, which would be added to the product of third<br />
element in V a1 multiplied by the r 13 value, and so on for all the corresponding elements in V a1<br />
and the first row of R. To obtain the second element of U a2 , the same process would be<br />
repeated i.e., the elements in V a1 are accumulatively multiplied by the 2nd row of R. The<br />
same process would be repeated for each row of R and the result would be a new vector<br />
U a2 . Then U a2 would be normalized to obtain V a2 . One would then compare V a1 and V a2 . If<br />
they are nearly identical, then convergence is said to have occurred (If convergence does<br />
not occur, one should go on using these trial vectors again and again till convergence<br />
occurs). Suppose the convergence occurs when we work out V a8 in which case V a7 will be<br />
taken as V a (the characteristic vector) which can be converted into loadings on the first<br />
principal component when we multiply the said vector (i.e., each element of V a ) by the<br />
square root of the number we obtain for normalizing U a8 .<br />
(iii) To obtain factor B, one seeks solutions for V b , and the actual factor loadings for second<br />
component factor, B. The same procedures are used as we had adopted for finding the first<br />
factor, except that one operates off the first residual matrix, R 1 rather than the original<br />
correlation matrix R (We operate on R 1 in just the same way as we did in case of centroid<br />
method stated earlier).<br />
(iv) This very procedure is repeated over and over again to obtain the successive PC factors<br />
(viz. C, D, etc.).