Does Extraction Method Matter?

So, now that we have gone through the effort of learning about the different extraction methods and figuring how to run them in SAS, what difference do they really make? Let’s see by comparing the results of five different extraction methods using the data sets summarized in Example data sets. We ran these models using the same syntax presented in the section above, just changing the METHOD option and eliminating the PR IORS option for all except the PAF and iterated PAF methods. We then organized the results into some nice summary tables to permit easy comparison.
Engineering data. As we mentioned above, this data usually gives a clear two-factor solution. Communality estimates for the engineering data across different extraction techniques presents the communality estimates for the initial extraction, and each of the subsequent extraction methods. As you can see, the communalities for the various items are relatively stable despite the relatively small sample size, rarely varying by more than 0.10 across all extraction methods for a particular item. Eigenvalues extracted for the engineering data across different extraction techniques presents the eigenvalues extracted by the various methods. We also see very similar eigenvalues produced across the various methods. This suggests that when basic assumptions are met and factor structure is clear, the extraction method might not matter much.
Table 2.2 Communality estimates for the engineering data across different extraction techniques
Variable:
Initial
ML
PAF
Iterated PAF
ULS
ALPHA
EngProbSolv1
.742
.712
.731
.728
.728
.733
EngProbSolv2
.695
.663
.675
.669
.669
.669
EngProbSolv3
.752
.766
.764
.768
.768
.769
EngProbSolv4
.792
.810
.805
.810
.810
.810
EngProbSolv5
.790
.807
.797
.799
.799
.796
EngProbSolv6
.766
.774
.768
.768
.768
.767
EngProbSolv7
.786
.778
.778
.775
.775
.774
EngProbSolv8
.666
.674
.670
.671
.671
.669
INTERESTeng1
.674
.666
.672
.669
.668
.668
INTERESTeng2
.802
.834
.825
.833
.833
.834
INTERESTeng3
.816
.846
.835
.840
.840
.839
INTERESTeng4
.806
.831
.816
.817
.817
.813
INTERESTeng5
.781
.781
.796
.800
.800
.805
INTERESTeng6
.739
.750
.751
.752
.752
.751
Table 2.3 Eigenvalues extracted for the engineering data across different extraction techniques
Factor
Initial
ML
PAF
Iterated PAF
ULS
ALPHA
1
7.411
7.359
7.411
7.417
7.417
7.415
2
3.271
3.335
3.271
3.282
3.282
3.285
3
.197
4
.070
5
.070
Note: The ML and ALPHA methods generally report weighted eigenvalues as a result of the weighted estimation process. For the purpose of this example, the unweighted eigenvalues are reported. Factors 6-14 were suppressed from the initial extraction.
Self-description questionnaire data. As in the previous analysis, you can see that analyses of the SDQ by various extraction methods produce relatively similar results regardless of the extraction method. The communalities extracted were similar, and the eigenvalues were also similar (presented in Communality estimates for the SDQ data across different extraction techniques and Eigenvalues extracted for the SDQ data across different extraction techniques). These results further support the conclusion that extraction method is of little importance when basic assumptions are met and factor structure is clear.
Table 2.4 Communality estimates for the SDQ data across different extraction techniques
Variable
Initial
ML
PAF
Iterated PAF
ULS
ALPHA
Eng1
.537
.619
.598
.623
.623
.622
Eng2
.581
.676
.637
.664
.664
.648
Eng3
.608
.722
.678
.723
.724
.722
Eng4
.447
.403
.436
.413
.413
.425
Math1
.704
.790
.760
.792
.792
.794
Math2
.674
.751
.721
.737
.737
.721
Math3
.700
.783
.762
.799
.800
.816
Math4
.393
.372
.386
.371
.371
.374
Par1
.455
.526
.500
.510
.510
.496
Par2
.406
.434
.445
.450
.450
.458
Par3
.572
.695
.640
.678
.678
.668
Par4
.408
.392
.426
.421
.420
.442
Par5
.477
.557
.525
.539
.539
.525
Table 2.5 Eigenvalues extracted for the SDQ data across different extraction techniques
Factor
Initial
ML1
PAF
Iterated PAF
ULS
ALPHA
1
3.625
3.399
3.625
3.689
3.689
3.622
2
2.158
2.446
2.158
2.226
2.226
2.259
3
1.731
1.874
1.731
1.804
1.804
1.829
4
.362
5
-.021
6
-.054
Note. The ML and ALPHA methods generally report eigenvalues of the weighted reduced correlation matrix as a result of the weighted estimation process. For the purpose of this example, the unweighted eigenvalues are reported. Factors 7-13 were suppressed from the initial extraction.
GDS data. The goal of the third analysis is to compare the results of various extraction techniques on data with less clarity of structure. Because this data is binary (0, 1 values only) it is likely that it does not meet the assumption of multivariate normality. If one takes the advice above seriously, iterated PAF or ULS should be used given the non-normal data. As you can see in Comparison of communalities across extraction methods, the iterated PAF and ULS methods yield nearly identical results. The remaining methods differ, with substantial differences in communalities extracted for several variables in the scale. (See highlighted rows.) Comparing ML to iterated PAF and ULS, it is clear that the recommendation to use iterated PAF or ULS when data is not multivariate normal should be seriously considered. There are also some items in Comparison of communalities across extraction methods that exhibited substantial discrepancies between ML and iterated PAF or ULS (for example, see GDS01 or GDS12).
Table 2.6 Comparison of communalities across extraction methods
Variable
Initial
ML
PAF
Iterated PAF
ULS
ALPHA
GDS01
.518
.880
.579
.690
.688
.553
GDS02
.297
.346
.346
.366
.366
.367
GDS03
.513
.560
.564
.579
.580
.558
GDS04
.408
.612
.479
.577
.577
.550
GDS05
.400
.424
.399
.396
.395
.398
GDS06
.369
.450
.421
.447
.446
.447
GDS07
.451
.543
.494
.521
.521
.436
GDS08
.272
.276
.320
.329
.329
.391
GDS09
.559
.689
.620
.672
.672
.629
GDS10
.410
.416
.411
.406
.406
.397
GDS11
.310
.364
.362
.371
.371
.349
GDS12
.320
.996
.413
.788
.830
.660
GDS13
.278
.428
.320
.389
.386
.314
GDS14
.286
.406
.363
.451
.453
.489
GDS15
.384
.409
.425
.430
.430
.470
GDS16
.534
.564
.569
.567
.567
.561
GDS17
.500
.552
.539
.547
.546
.531
GDS18
.290
.264
.306
.281
.282
.314
GDS19
.396
.422
.426
.419
.419
.411
GDS20
.336
.355
.387
.387
.387
.462
GDS21
.346
.417
.412
.433
.432
.433
GDS22
.413
.471
.467
.491
.491
.514
GDS23
.254
.254
.258
.252
.252
.264
GDS24
.260
.280
.290
.282
.282
.311
GDS25
.442
.451
.482
.473
.473
.482
GDS26
.375
.445
.432
.437
.436
.425
GDS27
.211
.214
.245
.239
.239
.260
GDS28
.300
.309
.365
.333
.328
.336
GDS29
.195
.162
.194
.167
.167
.219
GDS30
.277
.368
.341
.362
.363
.380
Note: The max iterations had to be increased for the iterated PAF method to converge on a solution.
The initial iterated PAF extraction did not converge on a solution within the default number of iterations (25). This can indicate there is a problem with the data or analysis that is causing the estimates to be volatile and not converge on a solution. In this instance, the maximum number of iterations (MAXITER) was increased and a solution was converged upon at iteration 60. Since the iterated PAF results mimic those of the ULS extraction, we can conclude that there was not a problem with the analysis or results, but that ULS extraction might reach a solution faster than iterated PAF extraction.
The eigenvalues extracted also vary dramatically across extraction methods, as you can see in Comparison of extracted eigenvalues for different extraction techniques. Once again, ML produces the most unexpected results, but iterated PAF and ULS results are probably the most reliable. Note that ML produced smaller eigenvalues for the first factor than the second factor, which is unusual but is merely a reflection of the relatively extreme loadings produced.
Table 2.7 Comparison of extracted eigenvalues for different extraction techniques
Factor
Initial
ML
PAF
Iterated PAF
ULS
ALPHA
1
7.257
2.171
7.257
7.325
7.325
7.296
2
1.459
5.898
1.459
1.565
1.567
1.539
3
1.044
1.415
1.044
1.122
1.126
1.115
4
.675
1.030
.675
.787
.796
.744
5
.548
.977
.548
.698
.703
.635
6
.485
.755
.485
.625
.626
.649
7
.435
.623
.435
.538
.538
.512
8
.327
.457
.327
.421
.421
.419
9
.301
10
.227
Note. The ML and ALPHA methods generally report eigenvalues of the weighted reduced correlation matrix as a result of the weighted estimation process. For the purpose of this example, the unweighted eigenvalues are reported. Factors 11-39 were suppressed from the initial extraction.
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset