Factor Analysis: Multiple Groups

Similar documents
Exploratory Factor Analysis

Applications of Structural Equation Modeling in Social Sciences Research

PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA

The Role of Nonlinear Factor-to-Indicator Relationships in Tests of Measurement Equivalence

Evaluating Goodness-of-Fit Indexes for Testing Measurement Invariance

Rens van de Schoot a b, Peter Lugtig a & Joop Hox a a Department of Methods and Statistics, Utrecht

How to report the percentage of explained common variance in exploratory factor analysis

Richard E. Zinbarg northwestern university, the family institute at northwestern university. William Revelle northwestern university

Factorial Invariance in Student Ratings of Instruction

Overview of Factor Analysis

Extending the debate between Spearman and Wilson 1929: When do single variables optimally reproduce the common part of the observed covariances?

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS

Least Squares Estimation

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Pragmatic Perspectives on the Measurement of Information Systems Service Quality

interpretation and implication of Keogh, Barnes, Joiner, and Littleton s paper Gender,

UNDERSTANDING ANALYSIS OF COVARIANCE (ANCOVA)

Chapter 1 Introduction. 1.1 Introduction

SEM Analysis of the Impact of Knowledge Management, Total Quality Management and Innovation on Organizational Performance

Multiple regression - Matrices

Handling attrition and non-response in longitudinal data

Introduction to Matrix Algebra

Goodness of fit assessment of item response theory models

UNDERSTANDING THE DEPENDENT-SAMPLES t TEST

Exploratory Factor Analysis Brian Habing - University of South Carolina - October 15, 2003

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models

The primary goal of this thesis was to understand how the spatial dependence of

Testing for the Equivalence of Factor Covariance and Mean Structures: The Issue of Partial Measurement In variance

Power and sample size in multilevel modeling

Introduction to Data Analysis in Hierarchical Linear Models

Introduction to Fixed Effects Methods

Business Statistics. Successful completion of Introductory and/or Intermediate Algebra courses is recommended before taking Business Statistics.

[This document contains corrections to a few typos that were found on the version available through the journal s web page]

It is important to bear in mind that one of the first three subscripts is redundant since k = i -j +3.

UNDERSTANDING THE TWO-WAY ANOVA

PSYCHOLOGY Vol. II - The Construction and Use of Psychological Tests and Measures - Bruno D. Zumbo, Michaela N. Gelin, Anita M.

Test Bias. As we have seen, psychological tests can be well-conceived and well-constructed, but

Construct Equivalence as an Approach to Replacing Validated Cognitive Ability Selection Tests

Algebra Academic Content Standards Grade Eight and Grade Nine Ohio. Grade Eight. Number, Number Sense and Operations Standard

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

MULTIPLE REGRESSION WITH CATEGORICAL DATA

MATHEMATICS AS THE CRITICAL FILTER: CURRICULAR EFFECTS ON GENDERED CAREER CHOICES

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

Introduction to Longitudinal Data Analysis

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Factor analysis. Angela Montanari

FACTOR ANALYSIS NASC

Association Between Variables

Unrestricted versus restricted factor analysis of multidimensional test items: some aspects of the problem and some suggestions

Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary

Linearly Independent Sets and Linearly Dependent Sets

THE DIMENSION OF A VECTOR SPACE

Data, Measurements, Features

Missing Data. A Typology Of Missing Data. Missing At Random Or Not Missing At Random

LAGUARDIA COMMUNITY COLLEGE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF MATHEMATICS, ENGINEERING, AND COMPUTER SCIENCE

Elementary Statistics

Marketing Mix Modelling and Big Data P. M Cain

Constructing a TpB Questionnaire: Conceptual and Methodological Considerations

Q FACTOR ANALYSIS (Q-METHODOLOGY) AS DATA ANALYSIS TECHNIQUE

How To Understand Multivariate Models

individualdifferences

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling

How To Find Out How Different Groups Of People Are Different

Clinical Study Design and Methods Terminology

Wooldridge, Introductory Econometrics, 4th ed. Chapter 7: Multiple regression analysis with qualitative information: Binary (or dummy) variables

Course Text. Required Computing Software. Course Description. Course Objectives. StraighterLine. Business Statistics

Application of discriminant analysis to predict the class of degree for graduating students in a university system

Statistiek II. John Nerbonne. October 1, Dept of Information Science

STATISTICA Formula Guide: Logistic Regression. Table of Contents

Multivariate Analysis of Variance. The general purpose of multivariate analysis of variance (MANOVA) is to determine

1 Overview and background

Introduction to Path Analysis

Statistics courses often teach the two-sample t-test, linear regression, and analysis of variance

How Students Interpret Literal Symbols in Algebra: A Conceptual Change Approach

Linking Climate, Job Satisfaction and Contextual Performance to Customer Experience

The Latent Variable Growth Model In Practice. Individual Development Over Time

Part III. Item-Level Analysis

Risk Decomposition of Investment Portfolios. Dan dibartolomeo Northfield Webinar January 2014

Factor Rotations in Factor Analyses.

Longitudinal Meta-analysis

Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016

Additional sources Compilation of sources:

Statistics. Measurement. Scales of Measurement 7/18/2012

Multivariate Analysis of Variance (MANOVA): I. Theory

CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

Determining Minimum Sample Sizes for Estimating Prediction Equations for College Freshman Grade Average

Analyzing Structural Equation Models With Missing Data

Psychology 7291, Multivariate Analysis, Spring SAS PROC FACTOR: Suggestions on Use

by the matrix A results in a vector which is a reflection of the given

Regression III: Advanced Methods

Office of Institutional Research & Planning

The Relationship between the Fundamental Attribution Bias, Relationship Quality, and Performance Appraisal

Measuring Line Edge Roughness: Fluctuations in Uncertainty

Contrast Coding in Multiple Regression Analysis: Strengths, Weaknesses, and Utility of Popular Coding Structures

REVIEWING THREE DECADES WORTH OF STATISTICAL ADVANCEMENTS IN INDUSTRIAL-ORGANIZATIONAL PSYCHOLOGICAL RESEARCH

An Introduction to Path Analysis. nach 3

Algebra 1 Course Information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Transcription:

Factor Analysis: Multiple Groups TODD D. LITTLE AND DAVID W. SLEGERS Volume 2, pp. 617 623 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S. Everitt & David C. Howell John Wiley & Sons, Ltd, Chichester, 2005

Factor Analysis: Multiple Groups Factor Analysis: Multiple Groups with Means The confirmatory factor analysis (CFA) (see Factor Analysis: Confirmatory) model is a very effective approach to modeling multivariate relationships across multiple groups. The CFA approach to factorial invariance has its antecedents in exploratory factor analysis. Cattell [4] developed a set of principles by which to judge the rotated solutions from two populations with the goal being simultaneous simple structure. Further advancements were made by Horst & Schaie [7] and culminated with work by Meredith [13] in which he gave methods for rotating solutions from different populations to achieve one best fitting factor pattern. Confirmatory factor analytic techniques have made exploratory methods of testing for invariance obsolete by allowing an exact structure to be hypothesized. The multiplegroup CFA approach is particularly useful when making cross-group comparisons because it allows for (a) simultaneous estimation of all parameters (including mean-level information) for all groups and (b) direct statistical comparisons of the estimated parameters across the groups. The theoretical basis for selecting groups can vary from nominal variables such as gender, race, clinical treatment group, or nationality to continuous variables that can be easily categorized such as age-group or grade level. When making comparisons across distinct groups, however, it is critical to determine that the constructs of interest have the same meaning in each group (i.e., they are said to be measurement equivalent, or have strong factorial invariance; see below). This condition is necessary in order to make meaningful comparisons across groups [1]. In order to determine measurement equivalence, the analyses should go beyond the standard covariance structures information of the traditional CFA model to also include the mean structure information [9, 14, 16, 21]. We refer to such integrated modeling as mean and covariance structure (MACS) modeling. MACS analyses are well suited to establish construct comparability (i.e., factorial invariance) and, at the same time, detect possible between-group differences because they allow: (a) simultaneous model fitting of an hypothesized factorial structure in two or more groups (i.e., the expected pattern of indicator-to-construct relations for both the intercepts and factor loadings, (b) tests of the cross-group equivalence of both intercepts and loadings, (c) corrections for measurement error whereby estimates of the latent constructs means and covariances are disattenuated (i.e., estimated as true and reliable values), and (d) strong tests of substantive hypotheses about possible cross-group differences on the constructs [11, 14]. The General Factor Model. To understand the logic and steps involved in multiple-group MACS modeling, we begin with the matrix algebra notations for the general factor model, which, for multiple populations g = 1, 2,...,G, is represented by: X g = τ g + g ξ g + δ g (1) E(X g ) = µ xg = τ g + g κ g (2) g = g g g + g (3) where x is a vector of observed or manifest indicators, ξ is a vector of latent constructs, τ is a vector of intercepts of the manifest indicators, is the factor pattern or loading matrix of the indicators, κ represents the means of the latent constructs, is the variance-covariance matrix of the latent constructs, and is a symmetric matrix with the variances of the error term, δ, along the diagonal and possible covariances among the residuals in the off diagonal. All of the parameter matrices are subscripted with a g to indicate that the parameters may take different values in each population. For the common factor model, we assume that the indicators (i.e., items, parcels, scales, responses, etc.) are continuous variables that are multivariate normal (see Catalogue of Probability Density Functions) in the population and the elements of have a mean of zero and are independent of the estimated elements in the other parameter matrices. In a MACS framework, there are six types of parameter estimate that can be evaluated for equivalence across groups. The first three components refer to the measurement level: (a), the unstandardized regression coefficients of the indicators on the latent constructs (the loadings of the indicators), (b) τ, the

2 Factor Analysis: Multiple Groups intercepts or means of the indicators, and (c), the residual variances of each indicator, which is the aggregate of the unique factor variance and the unreliable variance of an indicator. The other three types of parameter refer to the latent construct level: (d) κ, the mean of the latent constructs, (e) φ ii latent variances, and (f ) φ ij latent covariances or correlations [9, 12, 14]. Taxonomy of Invariance A key aspect of multiple-group MACS modeling is the ability to assess the degree of factorial invariance of the constructs across groups. Factorial invariance addresses whether the constructs measurement properties (i.e., the intercepts and loadings, which reflect the reliable components of the measurement space) are the same in two or more populations. This question is distinct from whether the latent aspects of the constructs are the same (e.g., the constructs mean levels or covariances). This latter question deals with particular substantive hypotheses about possible group differences on the constructs (i.e., the reliable and true properties of the constructs). The concept of invariance is typically thought of and described as a hierarchical sequence of invariance starting with the weakest form and working up to the strictest form. Although we will often discuss the modeling procedures in terms of two groups, the extension to three or more groups is straightforward (see e.g., [9]). Configural Invariance. The most basic form of factorial invariance is ensuring that the groups have the same basic factor structure. The groups should have the same number of latent constructs, the same number of manifest indicators, and the same pattern of fixed and freed (i.e., estimated) parameters. If these conditions are met, the groups are said to have configural invariance. As the weakest form of invariance, configural invariance only requires the same pattern of fixed and freed estimates among the manifest and latent variables, but does not require the coefficients be equal across groups. Weak Factorial Invariance. Although termed weak factorial invariance, this level of invariance is more restricted than configural invariance. Specifically, in addition to the requirement of having the same pattern of fixed and freed parameters across groups, the loadings are equated across groups. The manifest means and residual variances are free to vary. This condition is also referred to as pattern invariance [15] or metric invariance [6]. Because the factor variances are free to vary across groups, the factor loadings are, technically speaking, proportionally equivalent (i.e., weighted by the differences in latent variances). If weak factorial invariance is found to be untenable (see testing below) then only configural invariance holds across groups. Under this condition, one has little basis to suppose that the constructs are the same in each group and systematic comparisons of the constructs would be difficult to justify. If invariance of the loadings holds then one has a weak empirical basis to consider the constructs to be equivalent and would allow cross-group comparisons of the latent variances and covariances, but not the latent means. Strong Factorial Invariance. As Meredith [14] compellingly argued, any test of factorial invariance should include the manifest means weak factorial invariance is not a complete test of invariance. With strong factorial invariance, the loadings and the intercepts are equated (and like the variances of the constructs, the latent means are allowed to vary in the second and all subsequent groups). This strong form of factorial invariance, also referred to as scalar invariance [22], is required in order for individuals with the same ability in separate groups to have the same score on the instrument. With any less stringent condition, two individuals with the same true level of ability would not have the same expected value on the measure. This circumstance would be problematic because, for example, when comparing groups based on gender on a measure of mathematical ability one would want to ensure that a male and a female with the same level of ability would receive the same score. An important advantage of strong factorial invariance is that it establishes the measurement equivalence (or construct comparability) of the measures. In this case, constructs are defined in precisely the same operational manner in each group; as a result, they can be compared meaningfully and with quantitative precision. Measurement equivalence indicates that (a) the constructs are generalizable entities in each subpopulation, (b) sources of bias and error (e.g., cultural bias, translation errors, varying conditions of administration) are minimal, (c) subgroup differences

Factor Analysis: Multiple Groups 3 have not differentially affected the constructs underlying measurement characteristics (i.e., constructs are comparable because the indicators specific variances are independent of cultural influences after conditioning on the construct-defining common variance; [14]), and (d) between-group differences in the constructs mean, variance, and covariance relations are quantitative in nature (i.e., the nature of group differences can be assessed as mean-level, variance, and covariance or correlational effects) at the construct level. In other words, with strong factorial invariance, the broadest spectrum of hypotheses about the primary construct moments (means, variances, covariances, correlations) can be tested while simultaneously establishing measurement equivalence (i.e., two constructs can demonstrate different latent relations across subgroups, yet still be defined equivalently at the measurement level). Strict Factorial Invariance. With strict factorial invariance, all conditions are the same as for strong invariance but, in addition, the residual variances are equated across groups. This level of invariance is not required for making veridical cross-group comparisons because the residuals are where the aggregate of the true measurement error variance and the indicator-specific variance is represented. Here, the factors that influence unreliability are not typically expected to operate in an equivalent manner across the subgroups of interest. In addition, the residuals reflect the unique factors of the measured indicators (i.e., variance that is reliable but unique to the particular indicator). If the unique factors differ trivially with regard to subgroup influences, this violation of selection theorem [14] can be effectively tolerated, if sufficiently small, by allowing the residuals to vary across the subgroups. In other words, strong factorial invariance is less biasing than strict factorial invariance because, even though the degree of random error may be quite similar across groups, if it is not exactly equal, the nonequal portions of the random error are forced into other parameters of a given model, thereby introducing potential sources of bias. Moreover, in practical applications of cross-group research such as cross-cultural studies, some systematic bias (e.g., translation bias) may influence the reliable component of a given residual. Assuming these sources of bias and error are negligible (see testing below), they could be represented as unconstrained residual variance terms across groups in order to examine the theoretically meaningful common-variance components as unbiasedly as possible. Partial Invariance. Widaman and Reise [23] and others have also introduced the concept of partial invariance, which is the condition when a constraint of invariance is not warranted for one or a few of the loading parameters. When invariance is untenable, one may then attempt to determine which indicators contribute significantly to the misfit ([3] [5]). It is likely that only a few of the indicators deviate significantly across groups, giving rise to the condition known as partial invariance. When partial invariance is discovered there are a variety of ways to proceed. (a) One can leave the estimate in the model, but not constrain it to be invariant across groups and argue that the invariant indicators are sufficient to establish comparability of the constructs [23]; (b) one can argue that the differences between indicators are small enough that they would not make a substantive difference and proceed with invariance constraints in place [9]; (c) one could decide to reduce the number of indicators by only using indicators that are invariant across groups [16]; (d) one could conclude that because invariance cannot be attained that the instrument must be measuring different constructs across the multiple groups and, therefore, not use the instrument at all [16]. Milsap and Kwok [16] also describe a method to assess the severity of the violations of invariance by evaluating the sensitivity and specificity at various selection points. Selection Theorem Basis for Expecting Invariance. The loadings and intercepts of a constructs indicators can be expected to be invariant across groups under a basic tenet of selection theorem namely, conditional independence ([8, 14]; see also [18]). In particular, if subpopulation influences (i.e., the basis for selecting the groups) and the specific components (unique factors) of the construct s manifest indicators are independent when conditioned on the common construct components, then an invariant measurement space can be specified even under extreme selection conditions. When conditional independence between the indicators unique factors and the selection basis hold, the construct information (i.e., common variance) contains, or carries, information about subpopulation influences. This expectation is quite reasonable if one assumes that the subpopulations derive from a common population from which the subpopulations can be described as selected on the basis of one or

4 Factor Analysis: Multiple Groups more criteria (e.g., experimental treatment, economic affluence, degree of industrialization, degree of individualism etc.). This expectation is also reasonable if one assumes on the basis of a specific theoretical view that the constructs should exist in each assessed subpopulation and that the constructs indicators reflect generally equivalent domain representations. Because manifest indicators reflect both common and specific sources of variance, cross-group effects may influence not only the common construct-related variance of a set of indicators but also the specific variance of one or more of them [17]. Measurement equivalence will hold if these effects have influenced only the common-variance components of a set of construct indicators and not their unique-specific components [8, 14, 18]. If cross-group influences differentially and strongly affect the specific components of indicators, nonequivalence would emerge. Although measurement nonequivalence can be a meaningful analytic outcome, it disallows, when sufficiently strong, quantitative construct comparisons. Identification Constraints There are three methods of placing constraints on the model parameters in order to identify the constructs and model (see Identification). When a mean structure is used, the location must be identified in addition to the scale of the other estimated parameters. The first method to identification and scale setting is to fix a parameter in the latent model. For example, to set the scale for the location parameters, one can fix the latent factor mean, κ, to zero (or a nonzero value). Similarly, to set the scale for the variancecovariance and loading parameters one can fix the variances, φ ii to 1.0 (or any other nonzero value). The advantages of this approach are that the estimated latent means in each subsequent group are relative mean differences from the first group. Because this first group is fixed at zero, the significance of the latent mean estimates in the subsequent groups is the significance of the difference from the first group. Fixing the latent variances to 1.0 has the advantage of providing estimates of the associations among the latent constructs in correlational metric as opposed to an arbitrary covariance metric. The second common method is known as the marker-variable method. To set the location parameters, one element of τ is set to zero (or a nonzero value) for each construct. To set the scale, one element of λ is fixed to 1.0 (or any other nonzero value) for each construct. This method of identification is less desirable than the 1st and 3rd methods because the location and scale of the latent construct is determined arbitrarily on the basis of which indicator is chosen. Reise, Widaman, and Pugh [19] recommend that if one chooses this approach the marker variables should be supported by previous research or selected on the basis of strong theory. A third possible identification method is to constrain the sum of τ for each factor to zero [20]. For the scale identification, the λs for a factor should sum to p, the number of manifest variables. This method forces the mean and variance of the latent construct to be the weighted average of all of its indicators means and loadings. The method has the advantage of providing a nonarbitrary scale that can legitimately vary across constructs and groups. It would be feasible, in fact, to compare the differences in means of two different constructs if one was theoretically motivated to do so (see [20], for more details of this method). Testing for Measurement Invariance and Latent Construct Differences In conducting cross-group tests of equality, either a statistical or a modeling rationale can be used for evaluating the tenability of the cross-group restrictions [9]. With a statistical rationale, an equivalence test is conducted as a nested-model comparison between a model in which specific parameters are constrained to equality across groups and one in which these parameters (and all others) are freely estimated in all groups. The difference in χ2 between the two models is a test of the equality restrictions (with degrees of freedom equal to the difference in their degrees of freedom). If the test is nonsignificant then the statistical evidence indicates no cross-group differences between the equated parameters. If it is significant, then evidence of cross-group inequality exists. The other rationale is termed a modeling rationale [9]. Here, model constraints are evaluated using practical fit indices to determine the overall adequacy of a fitted model. This rationale is used for large models with numerous constrained parameters because the χ2 statistic is an overly sensitive index of model fit, particularly for large numbers of constraints and when estimated on large sample sizes (e.g., [10]).

Factor Analysis: Multiple Groups 5 From this viewpoint, if a model with numerous constraints evinces adequate levels of practical fit, then the set of constraints are reasonable approximations of the data. Both rationales could be used in testing the measurement level and the latent level parameters. Because these two levels represent distinctly and qualitatively different empirical and theoretical goals, however, their corresponding rationale could also be different. Specifically, testing measurement equivalence involves evaluating the general tenability of an imposed indicator-to-construct structure via overall model fit indices. Here, various sources of model misfit (random or systematic) may be deemed substantively trivial if model fit is acceptable (i.e., if the model provides a reasonable approximation of the data; [2, 9]). The conglomerate effects of these sources of misfit, when sufficiently small, can be depicted parsimoniously as residual variances and general lack of fit, with little or no loss to theoretical meaningfulness (i.e., the trade-off between empirical accuracy and theoretical parsimony; [11]). When compared to a non-invariance model, an invariance model differs substantially in interpretability and parsimony (i.e., fewer parameter estimates than a noninvariance model), and it provides the theoretical and mathematical basis for quantitative between-group comparisons. In contrast to the measurement level, the latent level reflects interpretable, error-free effects among constructs. Here, testing them for evidence of systematic differences (i.e., the hypothesis-testing phase of an analysis) is probably best done using a statistical rationale because it is a precise criteria for testing the specific theoretically driven questions about the constructs and because such substantive tests are typically narrower in scope (i.e., fewer parameters are involved). However, such tests should carefully consider issues such as error rate and effect size. Numerous examples of the application of MACS modeling can be found in the literature, however, Little [9] offers a detailed didactic discussion of the issues and steps involved when making cross-group comparisons (including the LISREL source code used to estimate the models and a detailed Figural representation). His data came from a cross-cultural study of personal agency beliefs about school performance that included 2493 boys and girls from Los Angeles, Moscow, Berlin, and Prague. Little conducted an 8- group MACS comparison of boys and girls across the four sociocultural settings. His analyses demonstrated that the constructs were measurement equivalent (i.e., had strong factorial invariance) across all groups indicating that the translation process did not unduly influence the measurement properties of the instrument. However, the constructs themselves revealed a number of theoretically meaningful differences, including striking differences in the mean levels and the variances across the groups, but no differences in the strength of association between the two primary constructs examined. Extensions to Longitudinal MACS Modeling The issues related to cross-group comparisons with MACS models are directly applicable to longitudinal MACS modeling. That is, establishing the measurement equivalence (strong metric invariance) of a construct s indicators over time is just as important as establishing their equivalence across subgroups. One additional component of longitudinal MACS modeling that needs to be addressed is the fact that the specific variances of the indicators of a construct will have some degree of association across time. Here, independence of the residuals is not assumed, but rather dependence of the unique factors is expected. In this regard, the a priori factor model, when fit across time, would specify and estimate all possible residual correlations of an indicator with itself across each measurement occasion. Summary MACS models are a powerful tool for cross-group and longitudinal comparisons. Because the means or intercepts of measured indicators are included explicitly in MACS models, they provide a very strong test of the validity of construct comparisons (i.e., measurement equivalence). Moreover, the form of the group- or time-related differences can be tested on many aspects of the constructs (i.e., means, variances, and covariances or correlations). As outlined here, the tenability of measurement equivalence (i.e., construct comparability) can be tested using model fit indices (i.e., the modeling rationale), whereas specific hypotheses about the nature of possible group differences on the constructs can be tested using precise statistical criteria. A measurement equivalent model is advantageous for three reasons: (a) it is theoretically very parsimonious and, thus, a reasonable a

6 Factor Analysis: Multiple Groups priori hypothesis to entertain, (b) it is empirically very parsimonious, requiring fewer estimates than a non-invariance model, and (c) it provides the mathematical and theoretical basis by which quantitative cross-group or cross-time comparisons can be conducted. In other words, strong factorial invariance indicates that constructs are fundamentally similar in each group or across time (i.e., comparable) and hypotheses about the nature of possible group- or time-related influences can be meaningfully tested on any of the constructs basic moments across time or across each group whether the groups are defined on the basis of culture, gender, or any other grouping criteria. References [1] Bollen, K.A. (1989). Structural Equations with Latent Variables, Wiley, New York. [2] Browne, M.W. & Cudeck, R. (1993). Alternative ways of assessing model fit, in Testing Structural Equation Models, K.A. Bollen & J.S. Long, eds, Sage Publications, Newbury Park, pp. 136 162. [3] Byrne, B.M., Shavelson, R.J. & Muthén, B. (1989). Testing for the equivalence of factor covariance and mean structures: the issue of partial measurement invariance, Psychological Bulletin 105, 456 466. [4] Cattell, R.B. (1944). Parallel proportional profiles and other principles for determining the choice of factors by rotation, Psychometrika 9, 267 283. [5] Cheung, G.W. & Rensvold, R.B. (1999). Testing factorial invariance across groups: a reconceptualization and proposed new method, Journal of Management 25, 1 27. [6] Horn, J.L. & McArdle, J.J. (1992). A practical and theoretical guide to measurement invariance in aging research, Experimental Aging Research 18, 117 144. [7] Horst, P. & Schaie, K.W. (1956). The multiple group method of factor analysis and rotation to a simple structure hypothesis, Journal of Experimental Education 24, 231 237. [8] Lawley, D.N. & Maxwell, A.E. (1971). Factor Analysis as a Statistical Method, 2nd Edition, Butterworth, London. [9] Little, T.D. (1997). Mean and covariance structures (MACS) analyses of cross-cultural data: practical and theoretical issues, Multivariate Behavioral Research 32, 53 76. [10] Marsh, H.W., Balla, J.R. & McDonald, R.P. (1988). Goodness-of-fit indexes in confirmatory factor analysis: the effect of sample size, Psychological Bulletin 103, 391 410. [11] McArdle, J.J. (1996). Current directions in structural factor analysis, Current Directions 5, 11 18. [12] McArdle, J.J. & Cattell, R.B. (1994). Structural equation models of factorial invariance in parallel proportional profiles and oblique confactor problems, Multivariate Behavioral Research 29, 63 113. [13] Meredith, W. (1964). Rotation to achieve factorial invariance, Psychometrika 29, 187 206. [14] Meredith, W. (1993). Measurement invariance, factor analysis and factorial invariance, Psychometrika 58, 525 543. [15] Millsap, R.E. (1997). Invariance in measurement and prediction: their relationship in the single-factor case, Psychological Methods 2, 248 260. [16] Millsap, R.E. & Kwok, O. (2004). Evaluating the impact of partial factorial invariance on selection in two populations, Psychological Methods 9, 93 115. [17] Mulaik, S.A. (1972). The Foundations of Factor Analysis, McGraw-Hill, New York. [18] Muthén, B.O. (1989). Factor structure in groups selected on observed scores, British Journal of Mathematical and Statistical Psychology 42, 81 90. [19] Reise, S.P., Widaman, K.F. & Pugh, R.H. (1995). Confirmatory factor analysis and item response theory: two approaches for exploring measurement invariance, Psychological Bulletin 114, 552 566. [20] Slegers, D.W. & Little, T.D. (in press). Evaluating contextual influences using multiple-group, longitudinal mean and covariance structures (MACS) methods, in Modeling contextual influences in longitudinal data, T.D. Little, J.A. Bovaird & J. Marquis, eds, Lawrence Erlbaum, Mahwah. [21] Sörbom, D. (1982). Structural equation models with structured means, in Systems Under Direct Observation, K.G. Jöreskog & H. Wold, eds, Praeger, New York, pp. 183 195. [22] Steenkamp, J.B. & Baumgartner, H. (1998). Assessing measurement invariance in cross-national consumer research, Journal of Consumer Research 25, 78 90. [23] Widaman, K.F. & Reise, S.P. (1997). Exploring the measurement invariance of psychological instruments: applications in the substance use domain, in The science of prevention: Methodological advances from alcohol and substance abuse research, K.J. Bryant & M. Windle et al., eds, American Psychological Association, Washington, pp. 281 324. Further Reading MacCallum, R.C., Browne, M.W. & Sugawara, H.M. (1996). Power analysis and determination of sample size for covariance structure modeling, Psychological Methods 1, 130 149. McGaw, B. & Jöreskog, K.G. (1971). Factorial invariance of ability measures in groups differing in intelligence and socioeconomic status, British Journal of Mathematical and Statistical Psychology 24, 154 168. (See also Structural Equation Modeling: Latent Growth Curve Analysis) TODD D. LITTLE AND DAVID W. SLEGERS