Upload
tushar-kant
View
248
Download
0
Embed Size (px)
DESCRIPTION
factor anaysis
Citation preview
CAPACITY PLANNING AS A COMPETITIVE STRATEGY
FACTOR ANALYSIS
Introduction:
Any time it is desirable to collect as much data as possible in a research to find out a result. Thus, we find to collect data on many variables. But after collecting the data, one will face a situation on how to analyze those many variables. As the number of variables increases the number of correlations will increase faster than that. Then the problem of comprehending those variables into a manageable number comes into picture. This task of data reduction or summarization of data is achieved by Factor Analysis.
Factor Analysis is a multivariate statistical technique, which is used for data reduction by identifying an underlying structure in the data. For example, consider that we have collected data on the causes of breakdowns in a bus depot. The causes become variables. Assume, the number of variables has been found to be 40 variables. Now, factor analysis can reduce these 40 variables into 5 variables or factors or components. These 5 factors are weighted components or factors of those 40 variables.
Purpose of factor Analysis:
It is used as a data reduction technique and at the same time we can maintain as much of the original information as possible. That is the variance in the original data (say 100%) can be explained by an optimum number of reduced variables to an extent (say around 80%). But when we want to explain 100% variance, then in most cases the number of original variables will be equal to the number of reduced variables.
Qualitative and Quantitative distinction can be made in a large set of data e.g. Price, quality, faster). Factor analysis is used to find an underlying structure in a set of data.
Factor analysis can be used to find the scoring & researcher can find the weights to corresponding variables. These weights along with responses can be used to arrive at the overall scores of each respondent.
Factor analysis can transform the data into variables that are uncorrected or mostly have no correlation between them. Then this data can be used for further analysis such as multiple regressions.
STEPS IN ARRIVING AT PRINCIPAL COMPONENTS/FACTORS:
In general the process of factor analysis can be divided into three major steps
i) Formulation of the data set.
ii) Estimation of correlation/covariance matrix
iii) Extraction and rotation of factors
Now, let us look at the above these steps in detail and the working Methodology. Here a single example is taken throughout this chapter to illustrate the various calculations involved.
I ) Formulation of the data set:
Data set is to be formulated in accordance to the objective of the research. The scale used in the variables must be an interval scale or ratio scale. Then appropriate sample size has to be taken. Generally (9) it is better to take a sample size of about 4 or 5 times the number of variables. Though it is not mandatory
The following example analyzes socioeconomic data provided by Harman (1976). The five variables represent total population, median school years, total employment, miscellaneous professional services, and median house value. Each observation represents one of twelve census tracts in the Los Angeles Standard Metropolitan Statistical Area.
Table 1
Total populationMedian school yearsTotal employmentMiscellaneous professional servicesMedian house value
570012.8250027025000
100010.96001010000
34008.81000109000
380013.6170014025000
400012.8160014025000
82008.326006012000
120011.44001016000
910011.533006014000
990012.5340018018000
960013.7360039025000
96009.633008012000
940011.4400010013000
ii) Formulation of Correlation or covariance matrix:
The data set in the above step is converted into a correlation or a covariance matrix. Here we will see how to form a correlation matrix in our example. A correlation matrix is the matrix showing how the variables are correlated and respond with each other. Thus the diagonal elements will be correlations of each variable with itself and the value will be 1. Since, factor analysis groups the variables which are correlated with each other, correlation matrix should not be an identify matrix. This is tested using Bartlett test of sphericity (discussed later in the chapter)
A =
mean(A) = 6242 11 2333 121 17000x =1.0e+003 *
S=x'*x=1.0e+008 *
std(A) = 1.0e+003 *
D-1/2 =
R=1/11*D-1/2*S*D-1/2 =
While using covariance matrix the method is entirely similar except that instead of formulating correlation matrix, covariance matrix of the variables is formulated.
1/11*x*x =
Before binding out the correlation matrix, we should standardize the observations to get scale invariant. Then for the standardized values, we should find the correlation matrix.
Method of Extraction
There are various methods of extracting factors from the correlation/covariance matrix. They are
1. Principal component analysis
2. Common factor analysis
a) Principal factor analysis
b) Maximum likelihood method
c) Alpha method
d) Image factoring method
e) Unweighted least square method
f) Generalized least square method
In our context we will discuss principal component analysis, principal factor analysis and the maximum likelihood method.
Principal component analysis
This method finds out factors or components which are a linear combination or weighted components of the variables in the data set. Mathematically, it can be represented as
PC1 = W11X1 + W12X2 + W13X3 + + W1pXp
Where,
PC1 = Principal component
W11, W12, W13 = weights of the respective variable
Here, it is seen that the first principal component accounts for the maximum variance in the data set. The variance accounted by these p components will be equal to the total variance in the data. Thus, an mth component an be represented
PCm = Wm1X1 + Wm2X2 + Wm3X3 + + WmpXp
So, if there are p variables there will be p components or factors.
Thus, the data set with p variables and n observations
X1 X2 Xp
1
2 3 . : : : :
: . . . .
n
np
Geometric Interpretation of Principal Components:
X2
Pc1
X1 X3
Assume that we have N observations on three variables. The observations are plotted on the three dimensions.
X1X2X3
1X11X12X13
2X21X22X23
: : : :
iXi1Xi2Xi3
: : : :
NXN1XN2XN3
X12X22X32The resultant figure will be spherical nature as shown above. The objective is to capture the sample variances through different Principal components. From the figure, Pc1 component captures most of the variances or It reflects the spread of the data. If the Pc1 passes through the ( X1, X2,X3 ) and it makes angles 2 and 3 with X1, X2 and X3 axes. The principal component orientation is determined completely by the cosine of the angles. Now the equivalent point for ( Xi1,Xi2,Xi3 ) on Pc1 namely Yi1.
Yi1 = cos 1 ( Xi1 - X1 ) + cos 2 ( Xi2 - X2 ) + cos 3 ( Xi3 - X3 )
Where (cos 1)2 + (cos 2)2 + (cos 3)2 = 1
Now it is possible to arrive at for each Triplet (Xi1,Xi2,Xi3 ) on Yi1 . The mean of Y1 values is zero.
_ 1 N
Yi = ------- Yi1.
N i=1
_ 1 3
1
Yi = ------- Yi1 = ----- [ Y11 + Y21 ]
3 i=1
2
1
= ------ [cos 1 ( X11 - X1 ) + cos 2 ( X12 - X2 ) + cos 3 ( X13 - X3 )
2
cos 1 ( X21 - X1 ) + cos 2 ( X22 - X2 ) + cos 3 ( X23 - X3 ) ]
1
= cos 1 [ ----- (( X11 + X21 ) -X1) ]
2
cos 1 ( X1 - X1 ) = 0.
Now the variance on the principal component is the variance of X1, X2 and X3 variable.
So the variance of the Principal Component
1 N
= ------- ( Yi1 Y1 )2 [ Y1 = 0 ]
N-1 i=1
1 N
= ------- Yi12
N-1 i=1
Differentiate this variance by the angles 1,2 and 3 and equate it to zero, because we want to maximize the variance.
[ cos 1, cos 2 , cos 3 ] ( constitute the Principal Eigen vector. This will be the Eigen vector of the maximum characteristic root, arrived at by the highest Eigen value of the variance- covariance matrix.
[ cos 1, cos 2 , cos 3 ]s cos 1
cos 2= max
cos 3
In this similar manner, we can exhaust the next highest major axis and name it Pc2 and proceed as above, we can arrive at the second highest variance and we can prove that
2 = [ cos 1, cos 2 , cos 3 ]s cos 1
cos 2
cos 3
where 1, 2 and 3 are the angles made by the second highest major axis makes angle with the original axis X1,X2 and X3 . This will be the second highest characteristic root of the variance, covariance matrix of the original data.
The same logic could be extended to arrive at further Principal Components.
The variance = 1 1 = variance = Pc12
Subject to 1 1 = 1.
To determine the coefficient and solve it by Legranjian method.
------- [ Pc12 + 1( 1 - 1 1 )]
1
= ------- [1 1 + 1( 1 - 1 1 )] = 0.
1
( - I ) 1 = 0
( - I ) 1 = 0 (
the value of 1 is calculated from | - I | = 0
1 - 1I 1 = 0
1 1 - 1I 11 = 0
1 1 = 1 = variance.
Similarly same approach may be extended to arrive at the second, thirdand *** Principal Eigen vector. But we should keep in mind that
ii = 1
i . j = 0
This property help us to understand that the principal components are orthogonal.
Is converted into a correlation or covariance matrix of size pp. Then further it is transformed into
Pc 1 Pc 2 Pc p
1 w11 w21 wp12 w12 w22 wp23 w13 w23 wp3. : : : :
: . . . .
p w1pw2p wpp
pp
Since our objective was to reduce the number of variables, the number of components to be retained; which account for max variance with least number; can be done by various methods, for eg. Screenplot. They are discussed in further sections:
Principal Components from correlation matrix:
The correlation matrix as found in step 2 is
For this correlation matrix, eigen values and corresponding eigen vectors are calculated from the equation
| R - (I | = 0
Here ( are the eigen values
The eigen values are
For each eigen value the corresponding eigen vectors are
Thus each principal component is one eigen vector.
The principal components are: -
PC1 = 0.34265X1+ 0.45244X2 + 0.39666X3 + 0.55014X4+ 0.4668X5PC2 = -0.60162X1+ 0.40638X2 - 0.54165X3 + 0.077608X4+ 0.41654X5
PC3 = -0.06062X1 - 0.6897X2 - 0.2467X3 + 0.66326X4+ 0.14094X5
PC4 = -0.2024X1+ 0.35308X2 - 0.02528X3 + 0.50141X4 - 0.7631X5
PC5 = 0.68992X1+ 0.17264X2 - 0.69842X3 - 0.00018X4 - 0.08009X5
The variance explained by each principal component is the percent of eigen values. The individual and cumulative variances explained by the different principal components are shown in the following table.
Total Variance Explained
Initial EigenvaluesExtraction Sums of Squared Loadings
ComponentTotal% of VarianceCumulative %Total% of VarianceCumulative %
12.873257.46457.4642.873257.46457.464
21.796535.9393.3941.796535.9393.394
30.21514.30297.696
40.09941.98799.683
50.01580.316100.000
Principal Components from covariance matrix
Consider the covariance matrix obtained in step 2
S =
similar to what was done in the previous sections, here eigen values and corresponding eigen vectors are calculated.
The eigen values obtained are
Eigen vectors are: -
Initial Eigenvalues(a)Extraction Sums of Squared Loadings
ComponentTotal% of VarianceCumulative %Total% of VarianceCumulative %
140589811.9775.26075.26040589811.9775.26075.260
213279399.5624.62299.88213279399.5624.62299.882
360785.510.11399.995
42835.145.25710-599.9951
50.571.05710-8100.000
Test of significance:
If the eigen values are nearly zero, then we can not retain the principal components arrived due to those eigen values. So we need to test weather the zero eigen values are significant or not. For this kind of test Bartlett (1947) developed a procedure to test the hypothesis. Let us see how this test is used for the variance-covariance input as well as correlation input.
Variance Co-variance Input
Bartletts approximate 2 value for the above testing is given below
In the p variable matrix, p-k eigen values have nearly zero or zero values then we should see weather it is better to retain k components.
l(j) = jth eigen value from S
q = p k
M = n k (2q + 1 + 2/q)
The degrees of freedom are (p k 1) (p k + 2)
For the given example,
p = 5, q = p-k = 2, k = 3, n = 12
M = 12 3 1/6(2*2 + 1 +2/2)= 11
|S| = 5.2689 1022, l n |S| = 54.62
= ln 40589811.97 + ln 13279399.56 + ln 60785.51 = 44.92
= (53932832.74 53929997.04 = 1417.85
q ln l = 14.514
2 = 11( 54.62 + 44.92 + 14.514) = 52.95,
d.f. = (5 3 1)(5 3 + 2) = 2
Correlation Input
If the correlation matrix is arrived by the standardized data matrix, then we should test whether the population correlation matrix is identity matrix. If we prove that it is not identity matrix, then we could conclude that it has some eigen values which will be significant. The test used for this purpose is known as Bartletts Sphericity test
HO : = I
The test statistic
With degrees of freedom p (p-1)
For the given example,
2 = [12 1 (2 5 + 5)] ln 0.0018 = 53.72
degrees of freedom = (5)(4) = 10
We can reject the null hypothesis that population correlation matrix is the identity matrix at the 0.01 significance level.
No. of Principal components to be retained
Variance covariance input
First identify the number of significant components using Bartletts test. We cannot have all the principal components. So we should have to retain only less number of principal components from the significantly proved components.
The number of components to be retained may be determined by the cumulative percentage of variation:
Where represents the cumulative percentage of variation
Correlation Input:
The eigen values lacks meaning Kaiser (1958)
In this method the principal component should represent more variance than the single variable & the eigen value should be greater than one. If it is less than one that means that the factor does not even represents the variation of one variable.
Scree Test
Cattell (1966) has proposed a method to return the number of principal components. This test is called scree test. In this test, after getting all the eigen values, they are arranged in ascending order. Then the eigen values are taken on the Y-axis and the factors on the X-axis. If you plot usually, there is a chance that you will get elbow shaped curve.
For the given example, we get the following curve.
Draw a straight line for the lowest eigen values as shown in the figure. Then include those eigen values whose values are above the straight line and also the highest values in the straight line. From the figure it is clear that we should keep three factors.
Horn Test
Horn (1965) has suggested a method to retain the no. of principal components. The procedure is as follows. First, arrange the eigen values in ascending order and plot the eigen values against the number of factors. Then develop m sets of np normally distributed random variate, then we may have the p eigen values for each set. Then find out the average eigen values for each component and plot another curve with the average eigen value against the factors. Both the curves may intersect at one point. We should draw the line parallel to the Y-axis and if that line touches the number of component axis then retain those components.
COMMON FACTOR ANALYTIC MODELThe task is to identify the inter dependency between the variables x1, x2, , xp and representing the inter dependency by the m factors. Here the observable variables follow multi normal distribution. In common factor analytic model. Each variable could be written is the format given below:
In a matrix form it could be written as
Where,
,
,
Here the variances are normally distributed with mean 0 and variance 1. Similarly the are normally distributed with mean 0 and variance Var (ei) =
Where is called unique or specific variance. In this process the matrix is formed as follows:
The variance of ith response variances can be written as
if the correlation matrix is used as input, otherwise
The above said relation could be expressed in matrix form as
This expression is used to find the communalities of the response.
PRINCIPLE FACTOR MODELSEstimation of Factor loading by Principal factor method
We are not going into the detail of the mathematics of principal factor method and maximum likelihood method derivations. But how these two methods work are explained by the flow chart form in a lucid manner.
Principal factor method
For the given example (table 1)
Rnew (Diagonal elements replaced by the communalities)
=
Eigen Vectors
c =
Eigen Values
d =
First factor is found out by multiplying the eigen vector corresponding to highest eigen value by the square root of eigen value i.e.
We get,
W=
0.6254
0.7136
0.7144
0.8791
0.7421
Residue=Rnew-W*W'=
Eigen vectors for Residue
c =
Eigen values for Residue
d =
Second factor is found out by multiplying the eigen vector corresponding to highest eigen value by the square root of eigen value i.e.
We get,
W=
-0.7663
0.5551
-0.6792
0.1586
0.5781
Summary calculations for the given example:
Table 2
VariablesFactorsCommunalitySpecific
variance
F1F2
X1X2X3X4X50.6254
0.7136
0.7144
0.8791
0.7421-0.7663
0.5551
-0.6792
0.1586
0.57810.98
0.82
0.97
0.80
0.880.02
0.18
0.03
0.20
0.12
Total variance%
Common variance%
Eigen value54.68
61.44
2.734434.32
38.56
1.71618911
There is no change in the communality value between the unrotated factors and orthogonal rotated factors. But in the case of oblique rotation, there will be change in communality values arrived due to oblique rotation as well as communality values arrived due to unrotated factors. The reason may be that the angles between the factors are not constant that results in different loading that leads to different communality.
Maximum likelihood method:
This is the initialization matrix.
S=
1.0000 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 1.0000 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 1.0000 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 1.0000 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 1.0000 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 1.0000
The single factor solution is obtained to form the residual matrix.
=
0.6370
0.5840
0.9590
0.9600
0.9290
0.9360
The residual matrix is obtained by the formula:
S1= S T
T=
0.4058 0.3720 0.6109 0.6115 0.5918 0.5962
0.3720 0.3411 0.5601 0.5606 0.5425 0.5466
0.6109 0.5601 0.9197 0.9206 0.8909 0.8976
0.6115 0.5606 0.9206 0.9216 0.8918 0.8986
0.5918 0.5425 0.8909 0.8918 0.8630 0.8695
0.5962 0.5466 0.8976 0.8986 0.8695 0.8761
S1=
0.5942 0.2120 0.0041 -0.0105 -0.0218 0.0038
0.2120 0.6589 0.0159 -0.0306 -0.0165 0.0084
0.0041 0.0159 0.0803 0.0194 -0.0159 -0.0196
-0.0105 -0.0306 0.0194 0.0784 -0.0148 -0.0126
-0.0218 -0.0165 -0.0159 -0.0148 0.1370 0.0545
0.0038 0.0084 -0.0196 -0.0126 0.0545 0.1239
Find the Eigen vector and Eigen values of S1
Eigen vectors of S1
0.0022 0.0237 0.0137 0.0331 -0.7589 0.6498
-0.0586 0.0041 0.0184 -0.0066 0.6494 0.7579
0.6848 0.0473 0.6862 -0.2394 0.0176 0.0188
-0.6905 -0.2669 0.6400 -0.1989 -0.0362 -0.0382
-0.1300 0.6133 0.3331 0.7029 0.0248 -0.0366
0.1843 -0.7415 0.0895 0.6387 0.0140 0.0097
0.3474 0.3264 0.4434 0.4400 0.4345 0.4402
Eigen values of S1
0.0571 0 0 0 0 0
0 0.0754 0 0 0 0
0 0 0.0886 0 0 0
0 0 0 0.1953 0 0
0 0 0 0 0.4128 0
0 0 0 0 0 0.8435
The scaled matrix is obtained by using the value:
C * W T * C * W = MaxW=
0.6498
0.7580
0.0189
-0.0383
-0.0366
0.0096
The transpose of W is found out
0.6498 0.7580 0.0189 -0.0383 -0.0366 0.0096
WTW = 1.0001
Find the Square root of ( Max/ WTW)
C= 0.9184 and C * W * C * W1= 0.8435 = Max
=
0.6370 0.5968
0.5840 0.6962
0.9590 0.0174
0.9600 -0.0352
0.9290 -0.0336
0.9360 0.0088
T = 0.6370 0.5840 0.9590 0.9600 0.9290 0.9360
0.5968 0.6962 0.0174 -0.0352 -0.0336 0.0088
Iteration 1
Now = diagonal of (S- T) T = 0.7619 0.7875 0.6213 0.5905 0.5717 0.6015
0.7875 0.8258 0.5722 0.5361 0.5191 0.5528
0.6213 0.5722 0.9200 0.9200 0.8903 0.8978
0.5905 0.5361 0.9200 0.9228 0.8930 0.8983
0.5717 0.5191 0.8903 0.8930 0.8642 0.8692
0.6015 0.5528 0.8978 0.8983 0.8692 0.8762
=
0.2381 0 0 0 0 0
0 0.1742 0 0 0 0
0 0 0.0800 0 0 0
0 0 0 0.0772 0 0
0 0 0 0 0.1358 0
0 0 0 0 0 0.1238
Find the eigen vector and eigen values of
Eigen vectors of
0 0 0 0 0 1
0 0 0 0 1 0
0 1 0 0 0 0
1 0 0 0 0 0
0 0 0 1 0 0
0 0 1 0 0 0
Eigen values of
0.0772 0 0 0 0 0
0 0.0800 0 0 0 0
0 0 0.1238 0 0 0
0 0 0 0.1358 0 0
0 0 0 0 0.1742 0
0 0 0 0 0 0.2381
Next we find the product of ( -1/2 * (S- ) * -1/2)
-1/2 = (eigen vectors) * diag( 1\square root of eigen values) * ( transpose of eigen vectors )
-1/2 =
3.5991 0 0 0 0 0
0 3.5355 0 0 0 0
0 0 2.8421 0 0 0
0 0 0 2.7136 0 0
0 0 0 0 2.3959 0
0 0 0 0 0 2.0494
S- =
0.7619 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 0.8258 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 0.9200 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 0.9228 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 0.8642 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 0.8762
-1/2 * (S- ) * -1/2 =
9.8693 7.4312 6.2908 5.8697 4.9152 4.4256
7.4312 10.3223 5.7878 5.0848 4.4556 4.0213
6.2908 5.7878 7.4313 7.2496 5.9582 5.1140
5.8697 5.0848 7.2496 6.7952 5.7018 4.9273
4.9152 4.4556 5.9582 5.7018 4.9608 4.5370
4.4256 4.0213 5.1140 4.9273 4.5370 3.6801
Find the eigen vectors and eigen values of this above matrix
Eigen vectors=
0.4624 0.3688 0.8056 0.0243 -0.0191 0.0144
0.4428 0.6901 -0.5694 -0.0110 -0.0218 -0.0534
0.4418 -0.3178 -0.1019 -0.4643 0.1724 0.6694
0.4154 -0.3704 -0.0467 -0.4105 -0.1218 -0.7104
0.3546 -0.3076 -0.0980 0.5604 -0.6554 0.1626
0.3109 -0.2344 -0.0681 0.5487 0.7246 -0.1328
Eigen values =
35.0152 0 0 0 0 0
0 5.5485 0 0 0 0
0 0 2.5090 0 0 0
0 0 0 0.4149 0 0
0 0 0 0 -0.2722 0
0 0 0 0 0 -0.1564
Similarly find the C value for Max and Max -1
C = 5.9174 and C= 2.3555
The new W matrix is
2.7362 0.8687
2.6202 1.6255
2.6143 -0.7486
2.4581 -0.8724
2.0983 -0.7246
1.8397 -0.5521
= (eigen vectors of ) * diag( square root of eigen values of ) * ( transpose of eigen vectors of )
=
0.2778 0 0 0 0 0
0 0.2828 0 0 0 0
0 0 0.3519 0 0 0
0 0 0 0.3685 0 0
0 0 0 0 0.4174 0
0 0 0 0 0 0.4880
Finally find which is the product of and W
=
0.7601 0.2413
0.7410 0.4597
0.9200 -0.2634
0.9058 -0.3215
0.8758 -0.3024
0.8978 -0.2694
This process is continued.
Iteration 2
T =
0.7601 0.7410 0.9200 0.9058 0.8758 0.8978
0.2413 0.4597 -0.2634 -0.3215 -0.3024 -0.2694
T =
0.6360 0.6742 0.6357 0.6109 0.5927 0.6174
0.6742 0.7604 0.5606 0.5234 0.5100 0.5414
0.6357 0.5606 0.9158 0.9180 0.8854 0.8969
0.6109 0.5234 0.9180 0.9238 0.8905 0.8998
0.5927 0.5100 0.8854 0.8906 0.8585 0.8678
0.6174 0.5414 0.8969 0.8998 0.8677 0.8786
= diag (S - T )
0.3640 0 0 0 0 0
0 0.2396 0 0 0 0
0 0 0.0842 0 0 0
0 0 0 0.0762 0 0
0 0 0 0 0.1415 0
0 0 0 0 0 0.1214
Find the eigen vector and eigen values of
Eigen vectors of
0 0 0 0 0 1
0 0 0 0 1 0
0 1 0 0 0 0
1 0 0 0 0 0
0 0 0 1 0 0
0 0 1 0 0 0
Eigen values of
0.0762 0 0 0 0 0
0 0.0842 0 0 0 0
0 0 0.1214 0 0 0
0 0 0 0.1415 0 0
0 0 0 0 0.2396 0
0 0 0 0 0 0.3640
Next we find the product of ( -1/2 * (S- ) * -1/2)
-1/2 = (eigen vectors) * diag( 1\square root of eigen values) * ( transpose of eigen vectors )
-1/2 =
3.6226 0 0 0 0 0
0 3.4462 0 0 0 0
0 0 2.8701 0 0 0
0 0 0 2.6584 0 0
0 0 0 0 2.0429 0
0 0 0 0 0 1.6575
S- =
0.6360 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 0.7604 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 0.9158 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 0.9238 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 0.8585 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 0.8786
-1/2 * (S- ) * -1/2 =
8.3464 7.2908 6.3943 5.7878 4.2183 3.6027
7.2908 9.0307 5.6972 4.8555 3.7032 3.1702
6.3943 5.6972 7.5439 7.1721 5.1304 4.1768
5.7878 4.8555 7.1721 6.5286 4.7629 3.9040
4.2183 3.7032 5.1304 4.7629 3.5829 3.1288
3.6027 3.1702 4.1768 3.9040 3.1288 2.4138
Find the eigen vectors and eigen values of this above matrix
Eigen vectors=
0.4720 0.3301 0.8154 0.0405 -0.0405 -0.0112
0.4493 0.7056 -0.5419 -0.0275 0.0627 -0.0442
0.4700 -0.3431 -0.1344 -0.4047 -0.5472 0.4243
0.4290 -0.3986 -0.0429 -0.3710 0.6187 -0.3672
0.3171 -0.2774 -0.1273 0.5847 -0.3888 -0.5596
0.2642 -0.1991 -0.0736 0.5952 0.4012 0.6082
Eigen values =
31.7649 0 0 0 0 0
0 4.5782 0 0 0 0
0 0 1.1586 0 0 0
0 0 0 0.3121 0 0
0 0 0 0 -0.1625 0
0 0 0 0 0 -0.2051
Similarly find the C value for Max and Max -1
C=5.6363 and C=2.1397
The new W matrix is
2.6603 0.7063
2.5324 1.5098
2.6491 -0.7341
2.4180 -0.8529
1.7873 -0.5936
1.4891 -0.4260
= (eigen vectors of ) * diag( square root of eigen values of ) * ( transpose of eigen vectors of )
=
0.2760 0 0 0 0 0
0 0.2902 0 0 0 0
0 0 0.3484 0 0 0
0 0 0 0.3762 0 0
0 0 0 0 0.4895 0
0 0 0 0 0 0.6033
Finally find which is the product of and W
=
0.7342 0.1949
0.7349 0.4381
0.9229 -0.2558
0.9097 -0.3209
0.8749 -0.2906
0.8984 -0.2570
Iteration 3:
T =
0.7342 0.7349 0.9229 0.9097 0.8749 0.8984
0.1949 0.4381 -0.2558 -0.3209 -0.2906 -0.2570
T =
0.5771 0.6250 0.6278 0.6054 0.5857 0.6095
0.6250 0.7320 0.5662 0.5279 0.5156 0.5476
0.6278 0.5662 0.9172 0.9217 0.8818 0.8949
0.6053 0.5279 0.9216 0.9305 0.8891 0.8997
0.5857 0.5157 0.8818 0.8891 0.8499 0.8607
0.6095 0.5476 0.8949 0.8997 0.8607 0.8731
= diag (S - T )
0.4229 0 0 0 0 0
0 0.2680 0 0 0 0
0 0 0.0828 0 0 0
0 0 0 0.0695 0 0
0 0 0 0 0.1501 0
0 0 0 0 0 0.1269
Find the eigen vector and eigen values of
Eigen vectors of
0 0 0 0 0 1
0 0 0 0 1 0
0 1 0 0 0 0
1 0 0 0 0 0
0 0 0 1 0 0
0 0 1 0 0 0
Eigen values of
0.0695 0 0 0 0 0
0 0.0828 0 0 0 0
0 0 0.1269 0 0 0
0 0 0 0.1501 0 0
0 0 0 0 0.2680 0
0 0 0 0 0 0.4229
Next we find the product of ( -1/2 * (S- ) * -1/2)
-1/2 = (eigen vectors) * diag( 1\square root of eigen values) * ( transpose of eigen vectors )
-1/2 =
3.7932 0 0 0 0 0
0 3.4752 0 0 0 0
0 0 2.8072 0 0 0
0 0 0 2.5811 0 0
0 0 0 0 1.9317 0
0 0 0 0 0 1.5377
S- =
0.5771 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 0.7320 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 0.9172 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 0.9305 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 0.8499 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 0.8731
-1/2 * (S- ) * -1/2 =
8.3035 7.6984 6.5487 5.8842 4.1766 3.4997
7.6984 8.8404 5.6192 4.7540 3.5311 2.9658
6.5487 5.6192 7.2279 6.8109 4.7448 3.7900
5.8842 4.7540 6.8109 6.1991 4.3726 3.5165
4.1766 3.5311 4.7448 4.3726 3.1714 2.7446
3.4997 2.9658 3.7900 3.5165 2.7446 2.0645
Find the eigen vectors and eigen values of this above matrix
Eigen vectors=
0.4928 -0.3041 -0.8065 0.0915 -0.0712 -0.0293
0.4603 -0.7017 0.5303 -0.0660 0.0991 -0.0200
0.4646 0.3580 0.1509 -0.3932 -0.6377 0.2682
0.4209 0.4163 0.0033 -0.3688 0.6950 -0.1747
0.3020 0.2753 0.1878 0.5730 -0.2087 -0.6526
0.2473 0.1946 0.1016 0.6069 0.2277 0.6859
Eigen values =
31.0110 0 0 0 0 0
0 4.2810 0 0 0 0
0 0 0.5792 0 0 0
0 0 0 0.2692 0 0
0 0 0 0 -0.1368 0
0 0 0 0 0 -0.1968
Similarly find the C value for Max and Max -1
C=5.5685 and C=2.0692
The new W matrix is
2.7442 -0.6292
2.5632 -1.4520
2.5871 0.7408
2.3438 0.8614
1.6817 0.5697
1.3771 0.4027
= (eigen vectors of ) * diag( square root of eigen values of ) * ( transpose of eigen vectors of )
=
0.2636 0 0 0 0 0
0 0.2877 0 0 0 0
0 0 0.3562 0 0 0
0 0 0 0.3874 0 0
0 0 0 0 0.5177 0
0 0 0 0 0 0.6503
Finally find which is the product of and W
=
0.7234 -0.1659
0.7374 -0.4177
0.9215 0.2639
0.9080 0.3337
0.8706 0.2949
0.8955 0.2619
Iteration 4:
T =
0.7234 0.7374 0.9215 0.9080 0.8706 0.8955
-0.1659 -0.4177 0.2639 0.3337 0.2949 0.2619
T =
0.5508 0.6027 0.6228 0.6015 0.5809 0.6043
0.6028 0.7183 0.5693 0.5302 0.5188 0.5510
0.6229 0.5693 0.9188 0.9248 0.8801 0.8943
0.6015 0.5302 0.9248 0.9358 0.8889 0.9005
0.5809 0.5188 0.8801 0.8889 0.8449 0.8569
0.6044 0.5510 0.8943 0.9005 0.8569 0.8705
= diag (S - T )
0.4492 0 0 0 0 0
0 0.2817 0 0 0 0
0 0 0.0812 0 0 0
0 0 0 0.0642 0 0
0 0 0 0 0.1551 0
0 0 0 0 0 0.1295
Find the eigen vector and eigen values of
Eigen vectors of
0 0 0 0 0 1
0 0 0 0 1 0
0 1 0 0 0 0
1 0 0 0 0 0
0 0 0 1 0 0
0 0 1 0 0 0
Eigen values =
0.0642 0 0 0 0 0
0 0.0812 0 0 0 0
0 0 0.1295 0 0 0
0 0 0 0.1551 0 0
0 0 0 0 0.2817 0
0 0 0 0 0 0.4492
Next we find the product of ( -1/2 * (S- ) * -1/2)
-1/2 = (eigen vectors) * diag( 1\square root of eigen values) * ( transpose of eigen vectors )
-1/2 =
3.9467 0 0 0 0 0
0 3.5093 0 0 0 0
0 0 2.7789 0 0 0
0 0 0 2.5392 0 0
0 0 0 0 1.8841 0
0 0 0 0 0 1.4920
S- =
0.5508 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 0.7183 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 0.9188 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 0.9358 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 0.8449 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 0.8705
-1/2 * (S- ) * -1/2 =
8.5795 8.0885 6.7450 6.0229 4.2385 3.5331
8.0885 8.8460 5.6171 4.7227 3.4778 2.9059
6.7450 5.6171 7.0952 6.6328 4.5813 3.6403
6.0229 4.7227 6.6328 6.0336 4.1957 3.3566
4.2385 3.4778 4.5813 4.1957 2.9993 2.5974
3.5331 2.9059 3.6403 3.3566 2.5974 1.9378
Find the eigen vectors and eigen values of this above matrix
Eigen vectors=
0.5086 -0.2921 -0.6952 0.3936 -0.1233 -0.0519
0.4659 -0.6952 0.4560 -0.2683 0.1401 0.0021
0.4588 0.3680 0.0345 -0.4357 -0.6497 0.2023
0.4130 0.4267 -0.1779 -0.3202 0.7109 -0.0891
0.2932 0.2759 0.4181 0.4379 -0.1188 -0.6762
0.2392 0.1946 0.3180 0.5375 0.1538 0.7009
Eigen values =
31.0718 0 0 0 0 0
0 4.1792 0 0 0 0
0 0 0.3142 0 0 0
0 0 0 0.2402 0 0
0 0 0 0 -0.1173 0
0 0 0 0 0 -0.1967
Similarly find the C value for Max and Max -1
C=5.5745 and C =2.0442
The new W matrix is
2.8352 -0.5971
2.5972 -1.4211
2.5576 0.7523
2.3023 0.8723
1.6344 0.5640
1.3334 0.3978
= (eigen vectors of ) * diag( square root of eigen values of ) * ( transpose of eigen vectors of )
=
0.2534 0 0 0 0 0
0 0.2850 0 0 0 0
0 0 0.3599 0 0 0
0 0 0 0.3938 0 0
0 0 0 0 0.5308 0
0 0 0 0 0 0.6702
Finally find which is the product of and W
=
0.7184 -0.1513
0.7402 -0.4050
0.9205 0.2708
0.9066 0.3435
0.8675 0.2994
0.8936 0.2666
Iteration 5:T =
0.7184 0.7402 0.9205 0.9066 0.8675 0.8936
-0.1513 -0.4050 0.2708 0.3435 0.2994 0.2666
T =
0.5390 0.5931 0.6204 0.5994 0.5779 0.6017
0.5930 0.7119 0.5717 0.5319 0.5209 0.5535
0.6203 0.5717 0.9206 0.9275 0.8796 0.8947
0.5994 0.5320 0.9276 0.9400 0.8894 0.9018
0.5779 0.5209 0.8796 0.8893 0.8422 0.8550
0.6017 0.5535 0.8948 0.9018 0.8551 0.8696
= diag (S - T )
0.4610 0 0 0 0 0
0 0.2881 0 0 0 0
0 0 0.0794 0 0 0
0 0 0 0.0600 0 0
0 0 0 0 0.1578 0
0 0 0 0 0 0.1304
Find the eigen vector and eigen values of
Eigen vectors of
0 0 0 0 0 1
0 0 0 0 1 0
0 1 0 0 0 0
1 0 0 0 0 0
0 0 0 1 0 0
0 0 1 0 0 0
Eigen values =
0.0600 0 0 0 0 0
0 0.0794 0 0 0 0
0 0 0.1304 0 0 0
0 0 0 0.1578 0 0
0 0 0 0 0.2881 0
0 0 0 0 0 0.4610
Next we find the product of ( -1/2 * (S- ) * -1/2)
-1/2 = (eigen vectors) * diag( 1\square root of eigen values) * ( transpose of eigen vectors )
-1/2 =
4.0825 0 0 0 0 0
0 3.5489 0 0 0 0
0 0 2.7692 0 0 0
0 0 0 2.5174 0 0
0 0 0 0 1.8631 0
0 0 0 0 0 1.4728
S- =
0.5390 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 0.7119 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 0.9206 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 0.9400 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 0.8422 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 0.8696
-1/2 * (S- ) * -1/2 =
8.9834 8.4612 6.9527 6.1766 4.3355 3.6076
8.4612 8.9662 5.6607 4.7350 3.4779 2.9009
6.9527 5.6607 7.0596 6.5529 4.5144 3.5809
6.1766 4.7350 6.5529 5.9571 4.1133 3.2850
4.3355 3.4779 4.5144 4.1133 2.9234 2.5354
3.6076 2.9009 3.5809 3.2850 2.5354 1.8863
Find the eigen vectors and eigen values of this above matrix
Eigen vectors=
-0.5211 -0.2847 0.2327 -0.7416 -0.1934 -0.0766
-0.4697 -0.6889 -0.1449 0.4968 0.1908 0.0234
-0.4535 0.3758 0.2500 0.4005 -0.6335 0.1701
-0.4060 0.4335 0.3655 0.0518 0.7136 -0.0421
-0.2872 0.2773 -0.6089 -0.0096 -0.0596 -0.6828
-0.2341 0.1958 -0.5983 -0.2000 0.1104 0.7048
Eigen values =
31.4847 0 0 0 0 0
0 4.1710 0 0 0 0
0 0 0.2635 0 0 0
0 0 0 0.1580 0 0
0 0 0 0 -0.1036 0
0 0 0 0 0 -0.1977
Similarly find the C value for Max and Max -1
C=5.6116 and C =2.0423
The new W matrix is
-2.9242 -0.5814
-2.6358 -1.4069
-2.5449 0.7675
-2.2783 0.8853
-1.6117 0.5663
-1.3137 0.3999
= (eigen vectors of ) * diag( square root of eigen values of ) * ( transpose of eigen vectors of )
=
0.2449 0 0 0 0 0
0 0.2818 0 0 0 0
0 0 0.3611 0 0 0
0 0 0 0.3972 0 0
0 0 0 0 0.5367 0
0 0 0 0 0 0.6790
Finally find which is the product of and W
=
-0.1424 -0.7161
-0.3965 -0.7428
0.2771 -0.9190
0.3516 -0.9049
0.3039 -0.8650
0.2715 -0.8920
Iteration 6:T =
-0.1424 -0.3965 0.2771 0.3516 0.3039 0.2715
-0.7161 -0.7428 -0.9190 -0.9049 -0.8650 -0.8920
T =
0.5331 0.5884 0.6187 0.5980 0.5762 0.6001
0.5884 0.7089 0.5727 0.5327 0.5220 0.5549
0.6186 0.5727 0.9213 0.9290 0.8791 0.8950
0.5980 0.5328 0.9291 0.9425 0.8896 0.9027
0.5761 0.5220 0.8792 0.8896 0.8406 0.8541
0.6001 0.5549 0.8950 0.9026 0.8541 0.8694
= diag (S - T )
0.4669 0 0 0 0 0
0 0.2911 0 0 0 0
0 0 0.0787 0 0 0
0 0 0 0.0575 0 0
0 0 0 0 0.1594 0
0 0 0 0 0 0.1306
Find the eigen vector and eigen values of
Eigen vectors of
0 0 0 0 0 1
0 0 0 0 1 0
0 1 0 0 0 0
1 0 0 0 0 0
0 0 0 1 0 0
0 0 1 0 0 0
Eigen values =
0.0575 0 0 0 0 0
0 0.0787 0 0 0 0
0 0 0.1306 0 0 0
0 0 0 0.1594 0 0
0 0 0 0 0.2911 0
0 0 0 0 0 0.4669
Next we find the product of ( -1/2 * (S- ) * -1/2)
-1/2 = (eigen vectors) * diag( 1\square root of eigen values) * ( transpose of eigen vectors )
-1/2 =
4.1703 0 0 0 0 0
0 3.5646 0 0 0 0
0 0 2.7671 0 0 0
0 0 0 2.5047 0 0
0 0 0 0 1.8534 0
0 0 0 0 0 1.4635
S- =
0.5331 0.5840 0.6150 0.6010 0.5700 0.6000
0.5840 0.7089 0.5760 0.5300 0.5260 0.5550
0.6150 0.5760 0.9213 0.9400 0.8750 0.8780
0.6010 0.5300 0.9400 0.9425 0.8770 0.8860
0.5700 0.5260 0.8750 0.8770 0.8406 0.9240
0.6000 0.5550 0.8780 0.8860 0.9240 0.8694
-1/2 * (S- ) * -1/2 =
9.2714 8.6814 7.0969 6.2777 4.4057 3.6619
8.6814 9.0075 5.6814 4.7320 3.4751 2.8953
7.0969 5.6814 7.0542 6.5149 4.4875 3.5556
6.2777 4.7320 6.5149 5.9128 4.0712 3.2477
4.4057 3.4751 4.4875 4.0712 2.8875 2.5063
3.6619 2.8953 3.5556 3.2477 2.5063 1.8621
Find the eigen vectors and eigen values of this above matrix
Eigen vectors=
-0.5290 -0.2817 0.1415 -0.7351 -0.2657 -0.0991
-0.4705 -0.6850 -0.0832 0.4919 0.2424 0.0416
-0.4507 0.3801 0.2815 0.4296 -0.6030 0.1577
-0.4018 0.4370 0.3797 -0.0584 0.7069 -0.0174
-0.2840 0.2781 -0.6067 0.0796 -0.0177 -0.6836
-0.2313 0.1965 -0.6177 -0.1529 0.0836 0.7043
Eigen values =
31.7739 0 0 0 0 0
0 4.1657 0 0 0 0
0 0 0.2578 0 0 0
0 0 0 0.0973 0 0
0 0 0 0 -0.1002 0
0 0 0 0 0 -0.1990
Similarly find the C value for Max and Max -1
C=0.1774 and C = 0.4899
The new W matrix is
-0.0938 -0.1380
-0.0835 -0.3356
-0.0800 0.1862
-0.0713 0.2141
-0.0504 0.1362
-0.0410 0.0963
= (eigen vectors of ) * diag( square root of eigen values of ) * ( transpose of eigen vectors of )
=
0.2398 0 0 0 0 0
0 0.2805 0 0 0 0
0 0 0.3614 0 0 0
0 0 0 0.3992 0 0
0 0 0 0 0.5395 0
0 0 0 0 0 0.6833
Finally find which is the product of and W
=
-0.0225 -0.0331
-0.0234 -0.0941
-0.0289 0.0673
-0.0285 0.0855
-0.0272 0.0735
-0.0280 0.0658
Factor Rotation: -
After extraction of the factors one needs to discriminate and say that these variables come under these factors. There should be (in an ideal case) no ambiguity or a single variable explained by more than one factor. One factor can explain the variance in data which was there by more than one variable, but the variance in one variable should be explained by one factor. This is an ideal situation, but in most cases this cannot be obtained. To achieve this end we will go for rotation of factors. These are basically two broad categories of rotation orthogonal rotation and oblique rotation. Under Orthogonal rotation we have methods / algorithms like varimax, quartimax and equimax methods. Some oblique rotational methods are oblimax, quartimin, covarimin, biquartimin and oblimin methods. In this chapter, varimax method and oblimax methods are discussed.
Once the attachment of a variable to a particular factor is a problem i.e. when it is difficult to interpret which variable is attached to which factor, the factors are subjected to rotation.
Time to ponder some of the features of factor loading.
We know
Consider the orthogonal matrix T we know that in the orthogonal matrix T = T-1. Rearrange the expression (1) by multiplying as follows:
This helps us to create more T matrix where the T and are having equivalence property. But the elements inside T is different from the . This provision helps us in the factor rotation.
This is done by multiplying the factor loadings by orthogonal matrix T given below: -
T=
This will assist in getting higher value for the factor loading and also that the small factor loading is becoming further negligible. The rotated loading will help us to assign the variables to the factors without much problem.
Thurstone (1945, Chap 14) has suggested simple structure principle:
1. Each row of matrix should contain at least one zero
2. Each column of should contain at least q zeros`(q=factor)3. Every pair of columns of should contain several responses whose loadings vanish in one column but not in the other.
4. If the number of factors q is four or more, every pair of columns of should contain a large number of responses with zero loadings in both columns.
5. Conversely, for every pair of columns of only a small number of responses should have non zero loadings in both columns.
In factor rotation we try to achieve the simple structure principle.
There are two types of rotation one is orthogonal rotation and another is oblique rotation.
Orthogonal Rotation - In orthogonal rotation, all the factors are rotated by the same angle. For example if there are two factors, then the angle by which the first factor is rotated is equal to the angle by which the second factor will be rotated as shown in the figure above. There are different methods available through which the orthogonal rotation is ensued. They are varimax, quartimax and equimax. Varimax method is explained in detail in the next section.
Oblique Rotation In oblique rotation, the angle of rotation of different factors are different i.e. in the case of two factors, the angle by which the first factor is rotated will be different from the angle by which the second factor is rotated. The methods available for the oblique rotation are oblimax, quartimin, biquartimin and oblimin.
Varimax rotation
Kaiser (1956, 1958, 1959)
Step 1: Get the normal loadings for the factors. Create pairs of factors. Select first pair.
Step 2: Calculate the angle (say,) at which the factors are rotated. Kaiser has shown that the angle must satisfy the relation.
The summation is carried out for p responses (variables).
Identify the quadrant for. This is done with the help of following table: -
Sign of denominatorSign of numerator
+-
+I. <
IV. -
EMBED Equation.3 <
-II.
EMBED Equation.3