18067_Chapter_32

download 18067_Chapter_32

of 23

Transcript of 18067_Chapter_32

  • 8/2/2019 18067_Chapter_32

    1/23

    488

    32BEST PRACTICES IN STRUCTURALEQUATION MODELING

    RA LPH O. MUELLER

    GREGORY R. HA NCOCK

    S

    tructural equation modeling (SEM) hasevolved into a mature and popularmethodology to investigate theory-derived

    structural/causal hypotheses. Indeed, with thecontinued development of SEM software pack-ages such as AMOS (Arbuckle, 2007), EQS(Bentler, 2006), LISREL (Jreskog & Srbom,2006), and Mplus (Muthn & Muthn, 2006),SEM has become the preeminent multivariatemethod of data analysis (Hershberger, 2003,pp. 4344).Yet, we believe that many practitionersstill have little, if any, formal SEM background,potentially leading to misapplications and pub-lications of questionable utility. Drawing on ourown experiences as authors and reviewers of

    SEM studies, as well as on existing guides forreporting SEM results (e.g., Boomsma, 2000;Hoyle & Panter, 1995; McDonald & Ho, 2002),we offer a collection of best practices guidelinesto those analysts and authors who contemplate

    using SEM to help answer their substantiveresearch questions. Throughout, we assume thatreaders have at least some familiarity with thegoals and language of SEM as covered in anyintroductory textbook (e.g., Byrne, 1998, 2001,2006; Kline, 2005; Loehlin, 2004; Mueller, 1996;Schumacker & Lomax, 2004). For those desiringeven more in-depth or advanced knowledge, werecommend Bollen (1989), Kaplan (2000), orHancock and Mueller (2006).

    SETTING THE STAGE

    The foundations of SEM are rooted in classical

    measured variable path analysis (e.g., Wright,1918) and confirmatory factor analysis (e.g.,Jreskog, 1966, 1967). From a purely statisticalperspective, traditional data analytical techniquessuch as the analysis of variance, the analysis of

    Authors Note: During the writing of this chapter, the first author was on sabbatical leave from The George Washington

    University and was partially supported by its Center for the Study of Language and Education and the Institute for Education

    Studies, both in the Graduate School of Education and Human Development. While on leave, he was visiting professor in the

    Department of Measurement, Statistics and Evaluation (EDMS) at the University of Maryland, College Park, and visiting

    scholar in its Center for Integrated Latent Variable Research (CILVR). He thanks the EDMS and CILVR faculty and staff for

    their hospitality, generosity, and collegiality. Portions of this chapter were adapted from a presentation by the authors at the

    2004 meeting of the American Educational Research Association in San Diego.

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 488

  • 8/2/2019 18067_Chapter_32

    2/23

    Structural Equation Modeling 489

    covariance, multiple linear regression, canonicalcorrelation, and exploratory factor analysisaswell as measured variable path and confirmatoryfactor analysiscan be regarded as special cases

    of SEM. However, classical path- and factor-ana-lytic techniques have historically emphasized anexplicit link to a theoretically conceptualizedunderlying causal model and hence are moststrongly identified with the more general SEMframework. Simply put, SEM defines a set ofdata analysis tools that allows for the testing oftheoretically derived and a priori specified causalhypotheses.

    Many contemporary treatments introduceSEM not just as a statistical technique but as a

    processinvolving several stages: (a) initial model

    conceptualization, (b) parameter identificationand estimation, (c) data-model fit assessment,and (d) potential model modification. As anystudy using SEM should address these fourstages (e.g., Mueller, 1997), we provide briefdescriptions here and subsequently use them asa framework for our best practices analysis illus-trations and publication guidelines.

    Initial Model Conceptualization

    The first stage of any SEM analysis should

    consist of developing a thorough understandingof, and justification for, the underlying theory ortheories that gave rise to the particular model(s)being investigated. In most of the traditionaland typical SEM applications, the operational-ized theories assume one of three forms:

    A measured variable path analysis(MVPA)

    model: hypothesized structural/causal

    relations among directly measured variables;

    the four-stage SEM process applied to MVPA

    models was illustrated in, for example,

    Hancock & Mueller, 2004.

    A confirmatory factor analysis(CFA) model:

    structural/causal relations between

    unobserved latent factors and their measured

    indicators; the four-stage SEM process applied

    to CFA models was illustrated in, for example,

    Mueller & Hancock, 2001.

    A latent variable path analysis(LVPA) model:

    structural/causal relations among latent

    factors. This type of SEM model is the focus

    in this chapter and constitutes a combination

    of the previous two. A distinction is made

    between the structural and the measurement

    portions of the model: While the former is

    concerned with causal relations among latent

    constructs and typically is the focus in LVPA

    studies, the latter specifies how these

    constructs are modeled using measuredindicator variables (i.e., a CFA model).

    More complex models (e.g., multisample,latent means, latent growth, multilevel, or mix-ture models) with their own specific recommen-dations certainly exist but are beyond thepresent scope. Regardless of model type, how-ever, a lack of consonance between model andunderlying theory will have negative repercus-sions for the entire SEM process. Hence, metic-ulous attention to theoretical detail cannot be

    overemphasized.

    Parameter Identification and Estimation

    A models hypothesized structural and non-structural relations can be expressed as popula-tion parameters that convey both magnitudeand sign of those relations. Before sample esti-mates of these parameters can be obtained, eachparameterand hence the whole modelmustbe shown to be identified; that is, it must be pos-sible to express each parameter as a function of

    the variances and covariances of the measuredvariables. Even though this is difficult and cum-bersome to demonstrate, fortunately, the identi-fication status of a model can often be assessedby comparing the total number of parametersto be estimated, t, with the number of unique(co)variances of measured variables,

    where p is the total number of measured vari-ables in the model. When t > u (i.e., whenattempting to estimate more parameters thanthere are unique variances and covariances), themodel is underidentified, and estimation ofsome (if not all) parameters is impossible. Onthe other hand, t u is a necessary but not suf-ficient condition for identification, and usuallyparameter estimation can commence: t = uimplies that the model is justidentified, whilet < u implies that it is overidentified (providedthat indeed all parameters are identified and anylatent variables in the system have been assignedan appropriate metric; see Note 4).

    u =pp + 1

    2,

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 489

  • 8/2/2019 18067_Chapter_32

    3/23

    490 BEST ADVANCED PRACTICES IN QUANTITATIVE METHODS

    SEM software packages offer a variety ofparameter estimation techniques for modelswhose identification can be established. Themost popular estimation method (and the

    default in most SEM software packages) is max-imum likelihood(ML), an iterative large-sampletechnique that assumes underlying multivariatenormality. Alternative techniques exist (e.g.,generalized least squares [GLS], asymptoticallydistribution free [ADF; Browne, 1984], androbust estimators [Satorra & Bentler, 1994]),some of which do not depend on a particularunderlying distribution of the data, but still, thevast majority of substantive studies use ML.

    Data-Model Fit Assessment

    A central issue addressed by SEM is how toassess the fit between observed data and thehypothesized model, ideally operationalized asan evaluation of the degree of discrepancybetween the true population covariance matrixand that implied by the models structural andnonstructural parameters. As the populationparameter values are seldom known, the differ-ence between an observed, sample-based covari-ance matrix and that implied by parameterestimates must serve to approximate the popula-tion discrepancy. For a justidentified model, theobserved data will fit the model perfectly: Thesystem of equations expressing each model param-eter as a function of the observed (co)variances isuniquely solvable; thus, the sample estimate ofthe model-implied covariance matrix will, bydefault, equal the sample estimate of the popula-tion covariance matrix. However, if a model isoveridentified, it is unlikely that these two matri-ces are equal as the system of equations (express-ing model parameters as functions of observedvariances and covariances) is solvable in morethan a single way.

    Abiding by a general desire for parsimony,overidentified models tend to be of more sub-stantive interest than justidentified ones becausethey represent simpler potential explanations ofthe observed associations. While data-model fitfor such models was initially conceived as a for-mal statistical test of the discrepancy between thetrue and model-implied covariance matrices (achi-square test with df = ut; Jreskog, 1966,1967), such a test now is often viewed as overlystrict given its power to detect even trivial devia-tions of a proposed model from reality. Hence,

    many alternative assessment strategies have

    emerged (for a now classic review, see Tanaka,1993) and continue to be developed. Data-modelfit indices for such assessment can be categorizedroughly into three broad classes (with recom-

    mended indices in italics):

    Absolute indices evaluate the overall dis-crepancy between observed and implied covari-ance matrices; fit improves as more parametersare added to the model and degrees of freedomdecrease: for example, the standardized rootmean square residual(SRMR), the chi-square test(recommended to be reported mostly for its his-torical significance), and the goodness-of-fitindex (GFI).

    Parsimonious indices evaluate the overalldiscrepancy between observed and impliedcovariance matrices while taking into account amodels complexity; fit improves as more param-eters are added to the model, as long as thoseparameters are making a useful contribution: forexample, the root mean square error of approxima-tion (RMSEA) with its associated confidenceinterval, the Akaike information criterion (AIC)for fit comparisons across nonnested models, andthe adjusted goodness-of-fit index (AGFI).

    Incremental indicesassess absolute or parsi-

    monious fit relative to a baseline model, usuallythe null model (a model that specifies no relationsamong measured variables): for example, thecomparative fit index(CFI), the normed fit index(NFI), and the nonnormed fit index (NNFI).

    If, after considering several indices, data-model fit is deemed acceptable (and judged bestcompared to competing models, if applicable),the model is retained as tenable, and individualparameters may be interpreted. If, however, evi-dence suggests unacceptable data-model fit, the

    next and often final stage in the SEM processis considered: modifying the model to improvefit in hopes of also improving the models corre-spondence to reality.

    Potential Model Modification

    In a strict sense, any hypothesized modelis, at best, only an approximation to reality; theremaining question is one of degree of that mis-specification. With regard to external specificationerrorswhen irrelevant variables were included inthe model or substantively important ones were left

    outremediation can only occur by respecifying

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 490

  • 8/2/2019 18067_Chapter_32

    4/23

    Structural Equation Modeling 491

    the model based on more relevant theory. On theother hand, internal specification errorswhenunimportant paths among variables were includedor when important paths were omittedcan

    potentially be diagnosed and remedied using Waldstatistics(predicted increase in chi-square if a previ-ously estimated parameter were fixed to someknown value, e.g., zero) and Lagrange multiplierstatistics (also referred to as modification indices;estimated decrease in chi-square if a previouslyfixed parameter were to be estimated). As thesetests recommendations are directly motivated bythe data and not by theoretical considerations, anyresulting respecifications must be viewed asexploratory in nature and might not lead to amodel that resembles reality any more closely than

    the one(s) initially conceptualized.

    BEST PRACTICES IN SEM DATAANALYSIS: A SET OF ILLUSTRATIONS

    Using the four-stage SEM process as a frame-work, we turn to an illustration of best practicesin the most common type of SEM analyses. Wechose to focus on a set of hypothesized modelsinvolving structural/causal relations amonglatent factors (i.e., LVPA models) to demonstrateour preference for using a two-phase approach(i.e., a measurement phase followed by a struc-tural phase) over a single-phase, all-in-oneanalysis. We conclude this section by illustratingthe statistical comparison of hierarchicallyrelated or nestedmodels (occurring, for example,when one models parameters are a proper sub-set of another models parameters) and address-ing the disattenuation (i.e., purification andstrengthening) of structural parameter esti-mates obtained from an LVPA when comparedwith those obtained from an analysis of the

    same overall structure but one that uses mea-sured variables only.

    Suppose an educational researcher is inter-ested in investigating the structural effects ofgirls reading and mathematics self-concept(Read-SC and Math-SC, respectively) on math-ematics proficiency (Math-Prof), as potentiallymediated by task-goal orientation (Task-Goal).More specifically, the investigator might havestrong theoretical reasons to believe that at leastone of three scenarios is tenable: In Model 1(Figure 32.1a), it is hypothesized that the effects

    of Read-SC and Math-SC on Math-Prof are

    both completely mediated by Task-Goal. InModel 2 (Figure 32.1b), only the effect of Read-SC on Math-Prof is completely mediated byTask-Goal, while Math-SC affects Math-Prof

    not only indirectly via Task-Goal but alsodirectly without other intervening variables.Finally, in Model 3 (Figure 32.1c), Read-SC andMath-SC are thought to affect Math-Profdirectly as well as indirectly via Task-Goal. Toillustrate the testing of the tenability of thesethree competing models, multivariate normaldata on three indicator variables for each of thefour constructs were simulated for a sample ofn = 1,000 ninth-grade girls. Table 32.1 describesthe 12 indicator variables in more detail, whileTable 32.2 contains relevant summary statistics.1

    At this point, it is possible and might seementirely appropriate to address the research ques-tions implied by the hypothesized models througha series of multiple linear regression (MLR) analy-ses. For example, for Model 2 in Figure 32.1b, twoseparate regressions could be conducted: (1) Anappropriate surrogate measure of Math-Profcould be regressed on proxy variables for Math-SC and Task-Goal, and (2) a suitable indicator ofTask-Goal could be regressed on proxies for Read-SC and Math-SC. If the researcher would chooseitems ReadSC3, MathSC3, TG1, and Proc from

    Table 32.1 as surrogates for their respective con-structs, MLR results would indicate that eventhough all hypothesized effects are statistically sig-nificantly different from zero, only small amountsof variance in the dependent variables TG1 andProc are explained by their respective predictorvariables (R2TG1 = 0.034, R

    2Proc = 0.26; see Table

    32.6 for the unstandardized and standardizedregression coefficients obtained from the twoMLR analyses2). As we will show through thecourse of the illustrations below, an appropriatelyconducted LVPA of the models in Figure 32.1 andthe data in Table 32.2 will greatly enhance the util-ity of the data to extract more meaningful resultsthat address the researchers key questions.

    SEM Notation

    As the three alternative structural modelsdepicted in Figure 32.1 are at the theoretical/latent construct level, we followed commonpractice and enclosed the four factors of Read-SC, Math-SC, Task-Goal, and Math-Prof inellipses/circles. On the other hand, a glanceahead at the operationalized model in Figure 32.2

    reveals that the now included measured variables

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 491

  • 8/2/2019 18067_Chapter_32

    5/23

    492 BEST ADVANCED PRACTICES IN QUANTITATIVE METHODS

    Read-SC

    (F1)

    (a) Model 1

    Task-Goal

    (F3)

    Math-Prof

    (F4)

    Math-SC

    (F2)

    cF1

    cF2

    cD3 cD4

    bF3F1

    bF4F3

    bF3F2

    cF2F1

    D3 D4

    Read-SC

    (F1)

    Task-Goal

    (F3)

    Math-Prof

    (F4)

    Math-SC

    (F2)

    cF1

    cF2

    cD3 cD4

    bF3F1

    bF4F3

    bF4F2bF3F2

    cF2F1

    D3 D4

    (b) Model 2

    Read-SC

    (F1)

    Task-Goal

    (F3)

    Math-Prof

    (F4)

    Math-SC

    (F2)

    cF1

    cF2

    cD3 cD4

    bF3F1

    bF4F3

    bF4F2

    bF4F1

    bF3F2

    cF2F1

    D3 D4

    (c) Model 3

    Figure 32.1 The theoretical models.

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 492

  • 8/2/2019 18067_Chapter_32

    6/23

    (items RSC1 to RSC3, MSC1 to MSC3, TG1 toTG3, Math, Prob, and Proc) are enclosed in rec-

    tangles/squares. Using the Bentler-WeeksVFED labeling convention (V for measuredVariable/item, F for latent Factor/construct, Efor Error/measured variable residual, D forDisturbance/latent factor residual), the latentand measured variables in the current modelsare labeled F1 through F4 and V1 through V12,respectively. The hypothesized presence orabsence of relations between variables in themodel is indicated by the presence or absence ofarrows in the corresponding path diagram:One-headed arrows signify direct structural or

    causal effects hypothesized from one variable toanother, while two-headed arrows denote

    hypothesized covariation and variation withoutstructural specificity. For example, for Model 1

    in Figure 32.1a, note (a) the hypothesizedcovariance between Read-SC and Math-SC andthe constructs depicted variances (two-headedarrows connect the factors to each other and tothemselves, given that a variables variance canbe thought of as a covariance of the variablewith itself), (b) the hypothesized structuraleffects of these two factors on Task-Goal (one-headed arrows lead from both to Task-Goal),but (c) the absence of such hypothesized directeffects on Math-Prof (there are no one-headedarrows directly leading from Read-SC and

    Math-SC to Math-Prof; the former two con-structs are hypothesized to affect the latter only

    Structural Equation Modeling 493

    Table 32.1 Indicator Variable/Item Description

    Construct Variable Label Item Scores

    Read-SC (F1)

    RSC1 (V1) Compared to others my age, 1 (false) to 6 (true)

    I am good at reading.

    RSC2 (V2) I get good grades in reading.

    RSC3 (V3) Work in reading class is easy for me.

    Math-SC (F2)

    MSC1 (V4) Compared to others my age, 1 (false) to 6 (true)

    I am good at math.

    MSC2 (V5) I get good grades in math.

    MSC3 (V6) Work in math class is easy for me.

    Task-Goal (F3)

    TG1 (V7) I like school work that Ill learn 1 (false) to 6 (true)

    from, even if I make a lot of mistakes.

    TG2 (V8) An important reason why I do

    my school work is because I like

    to learn new things.

    TG3 (V9) I like school work best when

    it really makes me think.

    Math-Prof (F4)

    Math (V10) Mathematics subtest scores

    of the Stanford AchievementTest 9

    Prob (V11) Problem Solving subtest

    scores of the Stanford

    Achievement Test 9

    Proc (V12) Procedure subtest scores

    of the Stanford Achievement

    Test 9

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 493

  • 8/2/2019 18067_Chapter_32

    7/23

    494 BEST ADVANCED PRACTICES IN QUANTITATIVE METHODS

    indirectly, mediated by Task-Goal). Finally,because variation in dependent variables usuallyis not fully explainable by the amount of variationor covariation in their specified causes, eachdependent variable has an associated residualterm. For example, in the operationalized modelin Figure 32.2, D3 and D4 denote the predictionerrors associated with the latent factors F3(Task-Goal) and F4 (Math-Prof), while E1through E3 indicate the residuals associated

    with the measured indicator variables (V1 toV3) of the latent construct Read-SC.

    For purposes of labeling structural and non-structural parameters associated with the connec-tions between measured and/or latent variables in apath diagram, we used the abcsystem3 (Hancock &Mueller, 2006, pp. 46). Structural effects from onevariable (measured or latent) to another are labeledbto,from, with the subscripts indicating the to andfromvariables (e.g.,in Figure 32.2,bF3F1 indicates the path

    Table 32.2 Correlations and Standard Deviations of Simulated Data

    READSC1 READSC2 READSC3 MATHSC1 MATHSC2 MATHSC3

    (V1) (V2) (V3) (V4) (V5) (V6)

    READSC1 1.000

    READSC2 0.499 1.000

    READSC3 0.398 0.483 1.000

    MATHSC1 0.206 0.148 0.123 1.000

    MATHSC2 0.150 0.244 0.095 0.668 1.000

    MATHSC3 0.121 0.091 0.308 0.633 0.641 1.000

    GOALS1 0.141 0.150 0.123 0.140 0.143 0.167

    GOALS2 0.123 0.151 0.134 0.163 0.180 0.145

    GOALS3 0.161 0.199 0.160 0.147 0.151 0.158

    SATMATH 0.049 0.007 0.003 0.556 0.539 0.521SATPROB 0.031 0.009 0.023 0.544 0.505 0.472

    SATPROC 0.025 0.029 0.006 0.513 0.483 0.480

    SD 1.273 1.353 1.285 1.396 1.308 1.300

    GOALS1 GOALS2 GOALS3 SATMATH SATPROB SATPROC

    (V7) (V8) (V9) (V10) (V11) (V12)

    READSC1

    READSC2

    READSC3

    MATHSC1MATHSC2

    MATHSC3

    GOALS1 1.000

    GOALS2 0.499 1.000

    GOALS3 0.433 0.514 1.000

    SATMATH 0.345 0.385 0.337 1.000

    SATPROB 0.304 0.359 0.281 0.738 1.000

    SATPROC 0.259 0.330 0.279 0.714 0.645 1.000

    SD 1.334 1.277 1.265 37.087 37.325 45.098

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 494

  • 8/2/2019 18067_Chapter_32

    8/23

    Structural Equation Modeling 495

    to F3 from F1, and bV2F1 denotes the path/factorloading to item V2 from factor F1). On the otherhand, variances and covariances are labeled by theletter c(e.g., in Figure 32.2, cF1 denotes the varianceof the latent construct F1, while cF2F1 represents thecovariance between the factors F2 and F1).

    All-in-One SEM AnalysisGenerally Not Recommended

    Although we generally do not recommendthe analytic strategy outlined in this section, itnevertheless will prove pedagogically instruc-tive and will motivate arguments in later sec-tions. With a hypothesized structure amonglatent constructs in place and associated mea-sured indicators selected, Model 1 in Figure 32.1acan be operationalized as illustrated in Figure32.2. This path diagram implies a set of 14structural equations, one for each dependentvariable: two equations from the structural por-tion of the model (i.e., the part that specifies thecausal structure among latent constructs) and12 equations from the measurement portion of

    the model (i.e., the part that links each of theindicator variables with the designated latentconstructs). Table 32.3 lists all 14 structuralequations and their associated endogenous(dependent) and exogenous (independent)variables that together specify the model inFigure 32.2 (variables are assumed to be mean-centered, thus eliminating the need for inter-cept terms; items V1, V4, V7, and V10 are usedas reference variables for their respective fac-tors, and thus their factor loadings are notfreeto be estimated butfixedto 1.0; also see Note 4).

    Though it might seem that the statisticalestimation of the unknown coefficients in thestructural equations (the b and c parameters)should be the focus at this stage of the analysis,a prior assessment of the data-model fit is moreessential as it allows for an overall judgmentabout whether the data fit the structure ashypothesized (indeed, should evidence materi-alize that the data do not fit the model, inter-pretations of individual parameter estimatesmight be useless). As can be verified from thepath diagram in Figure 32.2 by counting one- and

    cE7

    cE1

    cE2

    cE3

    cE4

    cE5

    cE6

    cE8 cE9 cE10 cE11 cE12

    cD3 cD4

    bV3F1 cF1

    cF2

    Proc

    (V12)

    Prob

    (V11)Math

    (V10)

    TG3

    (V9)

    TG2

    (V8)

    TG1

    (V7)

    E7 E8 E9 E10 E11 E12

    bV6F2

    bV8F3

    bV9F3 bV12F3

    bF4F3

    bF3F2

    bF3F1

    cF2F1

    bV11F4

    D3

    E4

    E1

    E2

    E3

    E5

    E6

    D4

    Math-Prof

    (F4)

    Task-Goal

    (F3)

    1

    1

    1

    1

    Math-SC

    (F2)

    Read-SC

    (F1)

    RSC3

    (V3)

    RSC2

    (V2)

    RSC1

    (V1)

    MSC1

    (V4)

    MSC2

    (V5)

    MSC3

    (V6)

    1 1 1 1 1 1

    1

    1

    1

    1

    1

    1

    1 1

    bV1F1

    bV5F2

    Figure 32.2 Initially operationalized Model 1.

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 495

  • 8/2/2019 18067_Chapter_32

    9/23

    496 BEST ADVANCED PRACTICES IN QUANTITATIVE METHODS

    Table 32.3 Structural Equations Implied by the Path Diagram in Figure 32.2

    Structural Portion

    Endogenous Variable Structural Equations Exogenous Variables a

    Task-Goal (F3) F3 = bF3F1 F1 + bF3F2 F2 + D3 Read-SC (F1)

    Math-SC (F2)

    Math-Prof (F4) F4 = bF4F3 F3 + D4 Task-Goal (F3)

    Measurement Portion

    Endogenous Variable Structural Equations Exogenous Variables b

    RSC1 (V1) V1 = (1)F1 + E1 Read-SC (F1)

    RSC2 (V2) V2 = bV2F1 F1 + E2

    RSC3 (V3) V3 = bV3F1 F1 + E3

    MSC1 (V4) V4 = (1)F2 + E4 Math-SC (F2)MSC2 (V5) V5 = bV5F2 F2 + E5

    MSC3 (V6) V6 = bV6F2 F2 + E6

    TG1 (V7) V7 = (1)F3 + E7 Task-Goal (F3)

    TG2 (V8) V8 = bV8F3 F3 + E8

    TG3 (V9) V9 = bV9F3 F3 + E9

    Math (V10) V10 = (1)F4 + E10 Math-Prof (F4)

    Prob (V11) V11 = bV11F4 F4 + E11

    Proc (V12) V12 = bV12F4 F4 + E12

    a. Residuals D, though technically independent, are not listed.b. Residuals E, though technically independent, are not listed.

    two-headed arrows labeled with bor csymbols,the model contains t= 28 parameters to be esti-mated:4 two variances of the independent latentconstructs and one covariance between them,two variances of residuals associated with thetwo dependent latent constructs, three pathcoefficients relating the latent constructs, eightfactor loadings, and 12 variances of residualsassociated with the measured variables. Further-

    more, the 12 measured variables in the modelproduce u = 12 (12 + 1)/2 = 78 unique variancesand covariances; the model is overidentified (t=28 < u = 78), and it is likely that some degree ofdata-model misfit exists (i.e., the observedcovariance matrix will likely differ, to somedegree, from that implied by the model). Toassess the degree of data-model misfit, variousfit indices can be obtained and then should becompared against established cutoff criteria(e.g., those empirically derived by Hu & Bentler,1999, and listed here in Table 32.4). Though

    here LISREL 8.8 (Jreskog & Srbom, 2006)

    was employed, running any of the availableSEM software packages will verify the followingdata-model fit results for the data in Table 32.2and the model in Figure 32.2 (because the dataare assumed multivariate normal, the maxi-mum likelihood estimation method was used):2 = 3624.59 (df= u t= 50,p < .001), SRMR =0.13, RMSEA = 0.20 with CI90: (0.19, 0.20), andCFI = 0.55.

    As is evident from comparing these resultswith the desired values in Table 32.4, the currentdata do not fit the proposed model; thus, it isnot appropriate to interpret any individual para-meter estimates as, on the whole, the model inFigure 32.2 should be rejected based on the cur-rent data. Now the researcher is faced with thequestion of what went wrong: (a) Is the sourceof the data-model misfit indeed primarily a flawin the underlying structural theory (Figure 32.1a),(b) can the misfit be attributed to misspecifica-tions in the measurement portion of the model

    with the hypothesized structure among latent

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 496

  • 8/2/2019 18067_Chapter_32

    10/23

    constructs actually having been specified cor-rectly, or (c) do misspecifications exist in boththe measurement and structural portions of themodel? To help address these questions and pre-vent potential confusion about the source ofobserved data-model misfit, we do not recom-

    mend that researchers conduct SEM analyses byinitially analyzing the structural and measure-ment portions of their model simultaneously,as was done here. Instead, analysts are urged tofollow a two-phase analysis process, as describednext.

    Two-Phase SEM AnalysisRecommended

    Usually, the primary reason for conceptualiz-ing LVPA models is to investigate the tenabilityof theoretical causal structures among latent

    variables. The main motivation for recommend-ing a two-phase process over an all-in-oneapproach is to initially separate a model into itsmeasurement and structural portions so thatmisspecifications in the former, if present, canbe realized and addressed first, before the struc-ture among latent constructs is assessed.5 Thisapproach will simplify the identification ofsources of data-model misfit and might also aidin the prevention of nonconvergence problemswith SEM software (i.e., when the iterative esti-mation algorithm cannot converge upon a

    viable solution for parameter estimates).

    Consider the path diagram in Figure 32.3a. Itis similar to the one depicted in Figure 32.2 as itinvolves the same measured and latent variablesbut differs in two important ways: Not only are

    Read-SC and Math-SC now explicitly connectedto Math-Prof, but all structural links amonglatent variables have been changed to nonstruc-tural relations (note in Figure 32.3a the two-headed arrows between all latent constructs thatare now labeled with csymbols). That is, latentconstructs are allowed to freely covary withoutan explicit causal structure among them. Inshort, Figure 32.3a represents a CFA model ofthe latent factors Read-SC, Math-SC, Task-Goal,and Math-Prof, using the measured variables inTable 32.1 as their respective effect indicators.6

    Measurement Phase. An analysis of the CFAmodel in Figure 32.3a constitutes the beginningof the measurement phaseof the proposed two-phase analysis process and produced the follow-ing data-model fit results: 2 = 3137.16 (df= 48,

    p < .001), SRMR = 0.062, RMSEA = 0.18 withCI90: (0.17, 0.18), and CFI = 0.61. These valuessignify a slight improvement over fit results forthe model in Figure 32.2. To the experiencedmodeler, this improvement was predictablegiven that the model in Figure 32.2 is more

    restrictive than, and a special case of, the CFAmodel in Figure 32.3a (with the paths fromRead-SC and Math-SC to Math-Prof fixed tozero); that is, the former model is nestedwithinthe latter,a topic more fully discussed in the nextsection. Irrespective of this minor improvement,however, why did the data-model fit remainunsatisfactory (as judged by the criteria listed inTable 32.4)? Beginning to analyze and addressthis misfit constitutes a move toward the fourthand final phase in the general SEM process,potential post hoc model modification.

    First, reconsider the list of items in Table 32.1.While all variables certainly seem to belong tothe latent factors they were selected to indicate,note that for the reading and mathematics self-concept factors, corresponding items are identicalexcept for one word: The word readingin itemsRSC1 through RSC3 was replaced by the wordmath to obtain items MSC1 through MSC3. Thus,it seems plausible that individuals responses tocorresponding reading and mathematics self-con-cept items are influenced by some of the same orrelated causes. In fact, the model in Figure 32.3a

    explicitly posits that two such related causes are

    Structural Equation Modeling 497

    Table 32.4 Target Values for Selected Fit Indicesto Retain a Model by Class

    Index Class

    Incremental Absolute Parsimonious

    NFI 0.90

    NNFI 0.95 GFI 0.90 AGFI 0.90

    CFI 0.95 SRMR 0.08 RMSEA 0.06

    Joint Criteria

    NNFI, CFI 0.96 and SRMR 0.09

    SRMR 0.09 and RMSEA 0.06

    SOURCE: Partially taken from Hu and Bentler (1999).

    NOTE: CFI = comparative fit index; NFI = normed fit index;

    NNFI = nonnormed fit index; GFI = goodness-of-fit index;

    AGFI = adjusted goodness-of-fit index; RMSEA = root mean

    square error of approximation; SRMR = standardized root

    mean square residual.

    32-Osborne (Best)-45409.qxd 10/9/2007 1:00 PM Page 497

  • 8/2/2019 18067_Chapter_32

    11/23

    498 BEST ADVANCED PRACTICES IN QUANTITATIVE METHODS

    the latent constructs Read-SC and Math-SC.However,as the specification of residual terms (E)indicates, responses to items are influenced bycauses other than the hypothesized latent con-

    structs. Those other, unspecified causes could alsobe associated. Thus, for example, the residualterms E1 and E4 might covary to some degree,particularly since both are associated with itemsthat differ by just one word. Based on similar the-oretical reasoning, a nonzero covariance mightexist between E2 and E5 and also between E3 andE6. In sum, it seems theoretically justifiable tomodify the CFA model in Figure 32.3a to allowresidual terms of corresponding reading andmathematics self-concept items to freely covary, asshown in Figure 32.3b. In fact, with enough fore-

    sight, these covariances probably should havebeen included in the initially hypothesized model.

    Second, as part of the analysis of the initial CFAmodel in Figure 32.3a, Lagrange multiplier (LM)statistics may be consulted for empirically basedmodel modification suggestions. These statisticsestimate the potential improvement in data-modelfit (as measured by the estimated decrease in chi-square) if a previously fixed parameter were to beestimated. Here, the three largest LM statistics were652.0, 567.8, and 541.7, associated with the fixedparameters cE5E2, cE4E1, and cE6E3, respectively.

    Compared with the overall chi-square value of3137.16, these estimated chi-square decreases seemsubstantial,7 foreshadowing a statistically signifi-cant improvement in data-model fit. Indeed, afterrespecifying the model accordingly (i.e., freeingcE5E2, cE4E1,and cE6E3; see Figure 32.3b) and reanalyz-ing the data, fit results for the modified CFA modelimproved dramatically: 2 = 108.04 (df= 45,p