Jump to content

Parallel analysis

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Staticshakedown (talk | contribs) at 21:20, 6 February 2019 (stub spacing). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Parallel analysis, also known as Horn's parallel analysis, is a statistical method used to determine the number of components to keep in a principal component analysis or factors to keep in an exploratory factor analysis. It is named after psychologist John L. Horn, who created the method in 1965.[1] The method compares the eigenvalues generated from the data matrix to the eigenvalues generated from a Monte-Carlo simulated matrix created from random data of the same size.[2]

Other methods of determining the number of factors or components to retain in an analysis include the scree plot or Kaiser rule.

Anton Formann provided both theoretical and empirical evidence that parallel analysis's application might not be appropriate in many cases since its performance is influenced by sample size, item discrimination, and type of correlation coefficient.[3]

See also

References

  1. ^ Horn, John L. (June 1965). "A rationale and test for the number of factors in factor analysis". Psychometrika. 30 (2): 179–185. doi:10.1007/bf02289447.
  2. ^ Mike Allen (11 April 2017). The SAGE Encyclopedia of Communication Research Methods. SAGE Publications. p. 518. ISBN 978-1-4833-8142-8.
  3. ^ Tran, U. S., & Formann, A. K. (2009). Performance of parallel analysis in retrieving unidimensionality in the presence of binary data. Educational and Psychological Measurement, 69, 50-61.