Beruflich Dokumente
Kultur Dokumente
DOI:10.1093/acprof:oso/9780195367621.003.0002
Keywords: structural equation modeling, SEM, social work research, SEM notation, equations
Page 1 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
We follow Bollen (1989) in making a critical distinction between the terms scale
and index. Note that this distinction is not made consistently in the literature!
The latent variable modeling that is the subject of this book specifically involves
scales, which in our conceptualization, are used to measure unobserved
phenomena that “cause” scores on multiple, correlated indicators (Bollen). An
underlying workplace climate will “cause” employees to respond in a generally
negative or positive way to a set of indicators on a workplace support scale. In
contrast, indicators of indices “cause” scores on the index and are not
necessarily highly correlated. Checking off items on a list (index or inventory) of
life stressors, for example, might lead to an individual’s high score on the index,
but experiencing the “death of a close family member,” “trouble with boss,” or
“pregnancy” are not necessarily or on average correlated or “caused” by some
underlying phenomenon (Holmes & Rahe, 1967). Scores on indices are not
driven by latent phenomena so are not of interest here.
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Page 3 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
In the SEM framework, the presence and nature of a latent variable such as
“indirect trauma exposure” or “perceived workplace support” is inferred from
relationships (correlations or covariances) among the scores for observed
variables chosen to measure it. Specifically, one starts with known information—
e.g., a covariance between two observed variables—and applies statistical
principles to estimate the relationship of each indicator to the hypothesized
latent variable. If we hypothesize the existence of the latent variable “ability,”
shown in Figure 2.1, for example, and we know from the questionnaire
responses of 200 subjects that the correlation between items Q1 and Q2 is 0.64,
we know (from measurement theory) that the product of the standardized paths
from “ability” to Q1 and Q2 equals 0.64 (DeVellis, 2003). If we assumed that the
two observed variables are influenced equally by the latent variable “ability,” we
would know that the path coefficients were both 0.80 (because 0.80 × 0.80 =
0.64). Squaring the path coefficients also indicates the
Page 4 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
In this discussion, we have illustrated an important property of SEM, that is, the
product of the standardized path coefficients (i.e., 0.80 and 0.80) from one latent
variable to two observed variables equals the correlation (i.e., 0.64) of the
observed variables. In Box 2.1, we provide a proof of the property, which was
developed by Spearman in 1904, marking the birth of SEM. In any SEM,
researchers have observed data, such as a known correlation of 0.64. The known
(or observed) data are used to estimate path coefficients, such as the two
coefficients reflecting the net influence of “ability” on Q1 and Q2. Of course, the
estimation becomes more complicated when there are multiple correlations or
covariances as input data, latent variable effects are not assumed to be the same
on all indicators, there are more than two indicators of a latent variable, and so
on. In more complicated models, in fact, more than one solution is possible—
more than one set of parameters might satisfy the multiple equations defining
the model. An important component of the analysis therefore becomes
determining which solution is the best. We will examine that issue more
thoroughly shortly.
Box 2-1 Proof of an SEM Property and a First Peek at SEM Notation
Page 5 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Q1 = λ 1 Ability + d1
Q2 = λ 2 Ability + d2 .
Assuming we work with standardized scores for all variables, then the
correlation ρ12 is simply the covariance of Q1 and Q2, or ρ12 = Cov(Q1, Q2).
Using the algebra of expectations, we can further write
Figure 2.2 presents a simple CFA model using common symbols. The model has
three latent variables: Risk1, Risk2, and Behavior. Latent variables are indicated
by circles or ovals. Because they are latent, by definition the three variables do
not exist in a dataset. They are hidden, unobservable, theoretical variables. In
the model, each is hypothesized to have three indicators. Risk1 represents some
risk phenomenon that influences (hence the one-way arrows) the scores
individuals in the database have on three observed variables, x1, x2 and x3. Often
latent variables have more than three indicators, especially when they represent
complex phenomena assessed with many scale items, or items on a
questionnaire. For example, 25 items assessing feelings of happiness, loneliness,
and sadness make up the Generalized Contentment Scale available at the
WALMYR website (http://www.walmyr.com/). It is also possible to have
Page 6 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
The common symbol for an observed variable in an SEM diagram (including CFA
models) is a square or rectangle. In Figure 2.2, x1, x2 and x3 are three
questionnaire items. Responses from the questionnaire items have been entered
into a database for analysis. The values may be numbers corresponding to
respondents’ answers to survey questions, or items on a rating scale, or values
coded from administrative, observational, or interview data. Observed variables
may also be recoded variables or composites based on other observed variables.
Like the Risk1 variable, Risk2 and Behavior are latent variables that are
hypothesized to “cause” the observed values of other questionnaire items (x4
through x9).
Page 7 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
The relationships among the latent and observed variables in Figure 2.2 can also
be expressed in equations that are similar to regression equations. The
equations relating latent variable Risk1 (ξ1, pronounced ksee) to x1, x2 and x3
are (p.24)
x 1 = λ 11 ξ 1 + δ 1
x 2 = λ 21 ξ 1 + δ 2
x 3 = λ 31 ξ 1 + δ 3
(Long, 1983).
The equations state that the score for an individual on any one observed variable
(x1, x2, x3) is the individual’s score on the latent variable times the factor loading
λ (lambda) of the observed variable on the latent variable, plus an error term δ
(delta). Note that the first subscript for a path coefficient (λ in these examples)
refers to the dependent variable in the equation—the variable to which an arrow
is pointing in the figure, or the variable on the left side of the equation. The
second subscript refers to the subscript of the independent variable.
The relationship between a latent factor (ξ) and one of its indicators is similar to
the regression relationship between a predictor, or independent variable, and a
dependent variable. The similarity reflects the fact that scores on the indicator
are “caused” by the latent variable. A critical difference, however, is that in
factor analysis, the predictor variable is unobserved, theoretical, or latent.
Without observed data in the dataset on the predictor, estimating its effects on
observed variables requires a different process than conventional regression
analysis (Long, 1983). It involves the use of matrix algebra and maximum
likelihood estimation, which will be discussed later. Still, the factor loading λ
that is obtained as an estimate of the strength of the effect of the latent variable
(the independent variable) on an indicator (dependent variable) is interpreted
the same as a regression coefficient—that is, a 1-unit change in the latent
variable is associated with a change of magnitude λ in the observed dependent
variable (Long). If variables are standardized, λ “is the expected shift in
standard deviation units of the dependent variable that is due to a one standard
deviation shift in the independent variable” (Bollen, 1989, p. 349).
Another difference between the latent variable equation and standard regression
equations is the lack of an intercept. Observed variables in SEM are treated as
deviations (or differences) from their means; in other words, instead of using the
raw scores that appear in a dataset, SEM software “centers” variables by
subtracting the mean from each score. This transformation has no effect on the
variances and covariances of (p.25) variables (the input data for SEM model
tests) but allows important simplifications of the equations used to estimate
models. Some of these simplifications were evident in the proof presented in Box
2.1. For further explanation, see Long (1983, pp. 22–23).
Page 8 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Page 9 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
SEM output provides estimates of the variances of the error terms for latent
variable indicators and indicates if they are statistically significantly different
from 0. Error variance is a summary measure of how much the error terms for a
sample on a predicted variable differ from the mean of those scores, which is
assumed to be 0. Larger error variances indicate that observed items are not
well explained by latent variables or may not be good measures of those latent
variables.
Page 10 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
(p.28) In summary,
measurement models include
latent factors and the
correlations among them,
observed indicators of those
factors, and error terms for
observed variables. Chapter 4
describes in more detail how to
specify confirmatory factor
models and interpret their
results.
Page 11 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Figure 2.4 presents a structural model based on Figure 2.2. Although the latent
variables and their relationships to indicator variables are still present, the
structural model has components that are different from the measurement
model. First, there are both single-headed and double-headed arrows among the
three latent variables in the model. Single headed arrows between two latent
variables indicate a hypothesized (p.29)
Page 12 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Note that the distinction between exogenous and endogenous variables is model
specific. Exogenous variables in one study may be (p.30) endogenous variables
in another study, or vice versa. Neighborhood cohesiveness might be an
exogenous predictor of the success of community organizing efforts in one
model, for example, but could be a dependent (endogenous) variable predicted
by community organizing efforts in another. Note also that to avoid confusion
between λ’s associated with exogenous (ξ) and endogenous (η) variables with
the same subscripts, we follow notation used by Bollen (1989) for models
containing both measurement and structural components. Instead of two
subscripts indicating the observed variable number and latent variable number,
respectively, λ’s are simply numbered consecutively with one subscript
throughout the model.
The SEM equation for regressing Behavior (η) on the two risk variables (ξ1, ξ2) is
η 1 = γ 11 ξ 1 + γ 12 ξ 2 + ζ 1
The equation states that the score for an individual on the latent behavior
variable (η1) is predicted by the individual’s score on the Risk1 latent variable
(ξ1) times the regression coefficient γ11 (gamma) plus the individual’s score on
the Risk2 latent variable (ξ2) times the regression coefficient γ12 plus the error
term ζ1 (zeta). ζ is structural error—the variance of Behavior that is unexplained
by its predictor variables. Structural error can also be thought of as the error of
prediction because, as in all regression analyses, variance in a dependent
variable (e.g., the endogenous Behavior variable) is likely to be influenced, or
predicted, by influences other than the variables included in a model. In other
words, we would not expect the risk and gender variables to predict Behavior
perfectly. Box 2.3 explains the difference between this type of error and the
measurement error we discussed in the previous section. Like the measurement
model equations, the equations predicting latent variable scores are similar to
regression equations, but with different notation and no intercepts. Latent
variable scores are also treated as deviations from their means.
Page 13 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
η 1 = γ 11 ξ 1 + γ 12 ξ 2 + γ 13 ξ 3 + ζ 1
Page 14 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
In this equation, Gender (ξ13) has been added as the third predictor of Behavior
(η1). γ13 is the regression coefficient representing the effect of Gender on
Behavior scores. Including the gender variable in Figure 2.4 illustrates how
structural models in SEM can include a combination of latent and observed
independent (and dependent) variables.
(p.32) The absence of double-headed arrows between Gender and the risk
variables signifies that the correlations between Gender and risk are expected to
be 0. It is important to remember that any possible path between two variables
in an SEM model that is not explicitly pictured represents the hypothesis that
the value of the directional or nondirectional relationship is 0. In the current
example, Risk1 and Risk2 might be latent measures of neighborhood
disorganization and danger, which we would not expect to be correlated with
gender.
y1=λ7η1+ε1
y2=λ8η1+ε2
y3=λ9η1+ε3.
All endogenous variables in a model are predicted (imperfectly) by one or more
other variables. Therefore, they all have associated error terms. If the predicted
variables are observed indicators of the measurement part of a model, the error
terms represent measurement error. If the variables are substantive dependent
variables (either latent or observed) being regressed on predictors, the error
terms represent structural errors.
Figure 2.5 presents a slightly different structural model. Risk1, Risk2, δ1 through
δ6, ε1 through ε4 and ζ1 and ζ2 are exogenous variables. Behavior, Parenting,
Par10 (y4), x1 through x6, and y1 through y3 are endogenous variables. There are
two structural errors, ζ1 and ζ2, and 10 measurement errors, δ1 through δ6, and
ε1 through ε4.
In Figure 2.5, Parenting is a new latent variable with one indicator (Par10). We
can imagine that the Parenting variable is an observed composite—the sum of
responses to 10 items on a parenting scale. Modeled as it is, Parenting is a
second endogenous latent variable whose value is equal to its one observed
indicator, which may or may not be (p.33)
Page 15 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
η 2 = γ 22 ξ 2 + ζ 2 .
Behavior (η1) is now predicted directly by the exogenous variable Risk1 (ξ1, with
a γ path), and by the endogenous Parenting variable with a β (beta) path.
Because there is no direct path from Risk2 to (p.34) Behavior, ξ2 does not
appear in the equation predicting Behavior (even though Risk2 has an indirect
effect on Behavior):
η 1 = β 12 η 2 + γ 11 ξ 1 + ζ 1 .
In summary, structural models in SEM models with latent variables have
measurement components and structural components. The structural paths
hypothesize substantive relationships among variables. Paths from exogenous (ξ)
to endogenous (η) latent variables are γ paths. Paths from endogenous to
endogenous latent variables are β paths. Observed indicators of exogenous
variables are “x” variables and have error terms labeled δ. Observed indicators
of endogenous variables are “y” variables and have measurement error terms
labeled ε. Structural errors, or errors of prediction, are designated with the
symbol ζ.
Page 16 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
(p.35) After data, a model, and an estimation procedure have been selected,
the SEM program iteratively generates estimates for parameters in the model,
which means the program continues to make and refine estimates that are
consistent with the input covariance matrix until no improvements can be made.
In a measurement model, the parameters to be estimated are the factor
loadings, latent variable variances and covariances, and measurement error
terms. In a general structural model, estimates of regression paths among latent
variables and structural error variances are also generated. A simplified version
of how the estimation process occurs was presented in the discussion of Figure
2.1. In reality, most models contain many parameters to be estimated, so the
program must attempt simultaneously to find estimates consistent with
numerous criteria, not just one initial covariance.
Page 17 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
What does it mean to say “no improvements” can be made in a model? The
determination of what are the best obtainable estimates is based on the
minimization of a function that the SEM program uses to compare the original
covariance matrix of the observed variables and a new covariance matrix that is
implied by the specified model and the estimates generated in the estimation
procedure. The new matrix is generated taking into account the constraints
imposed by the model specified by the user. For example, in Figure 2.5, a
moderate to strong covariance between observed variables x1 and x2 is
suggested by their common relationship to Risk1. In contrast, the model
suggests that the covariance between x1 and y1 is much smaller and occurs only
through the relationship of each observed variable with its latent variable. The
goal is to obtain an implied matrix that is as close to the original covariance
matrix as possible. The minimization function basically assesses how close each
element in the original covariance matrix is to its corresponding element in the
implied covariance matrix generated by each set of estimates tried. We will
return to this concept frequently because it is so key to understanding SEM.
Before we can go much further with this discussion of testing structural
equation models, we need to examine the numerous roles that matrices play in
SEM.
Matrices in SEM
A matrix is set of elements (i.e., numbers, values, or quantities) organized in
rows and columns. Most social work researchers are familiar (p.36) with
matrices. An Excel spreadsheet summarizing incoming students’ test scores,
grades, and demographics; a grading grid; and a proposal budget are just some
examples of matrices. The simplest matrix is one number, or a scalar. Other
simple matrices are vectors, which comprise only a row or column of numbers.
Correlation matrices are commonly encountered in social work research. They
summarize raw data collected from or about individuals and vary in size based
on the number of variables included. Correlation matrices have the same
number of rows and columns—one for each variable. Matrices can be multiplied,
divided, added, subtracted, inverted, transposed, and otherwise manipulated
following rules of matrix algebra.
Page 18 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Matrices are used in multiple ways in SEM analyses. Analyses rely, for example,
on data in the covariance or correlation matrices that summarize values in a raw
dataset. Also, all specified measurement and structural models with latent
variables are translated by SEM software into between three and eight matrices
(some of which may be vectors or scalars). The matrices are then manipulated
based on established proofs from matrix algebra and the algebra of expectations
to generate estimates of unknown parameters. Because matrices have known
properties and the outcomes of different operations on matrices (e.g., adding or
multiplying them together) are known, they provide a shortcut way—that is, a
faster, easier, less computationally demanding way—to accomplish the goals of
SEM analyses. As stated earlier, matrices are also the basis of the fundamental
SEM method of evaluating the quality of a model—comparing the original input
matrix to the model-implied matrix of covariances. More about each of these
roles of matrices in SEM is presented below. A full explanation of matrix algebra
is beyond the scope of this book. Bollen (1989) provides a useful summary for
interested readers. Long (1983) discusses matrix algebra as it applies to CFA.
x 1 = λ 11 ξ 1 + δ 1
x 2 = λ 21 ξ 1 + δ 2
x 3 = λ 31 ξ 1 + δ 3 .
The equations state that the observed scores of each x in the dataset are
predicted by a score on the latent factor (ξ1, Risk1) times a factor loading (λ)
plus error (δ). We can add similar equations for the rest of the observed
variables, which load on Risk2 (ξ2) and Behavior (ξ3) in the factor model in
Figure 2.2:
x 4 = λ 42 ξ 2 + δ 4
x 5 = λ 52 ξ 2 + δ 5
x 6 = λ 62 ξ 2 + δ 6
x 7 = λ 73 ξ 3 + δ 7
Page 19 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
x 8 = λ 83 ξ 3 + δ 8
x 9 = λ 93 ξ 3 + δ 9 .
All of these relationships can also be compactly expressed in the following
equation:
x=Λxξ+δ
where Λ (capital λ) is the matrix of λ’s, or factor loadings relating latent
variables to observed variables. The equation states more generally (p.38) that
the vector of values for a variable x in a raw dataset is a product of the variable’s
factor loading (Λ) on the latent variable (ξ) and the vector of scores for cases on
that latent variable, plus a vector of error terms.
The matrix format corresponding to both the detailed and compact equations is
[ x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 ] [ λ 11 0 0 λ 21 0 0 λ 31 0 0 0 λ 42 0 0 λ 52 0 0 λ 62 0 0 0
λ 73 0 0 λ 83 0 0 λ 93 ] [ ξ 1 ξ 2 ξ 3 ] + [ δ 1 δ 2 δ 3 δ 4 δ 5 δ 6 δ 7 δ 8 δ 9 ]
.
Brackets are used to enclose matrices. Mathematical symbols indicate the
operations specified for the matrices. The juxtaposition of the Λ and ξ matrices
indicate that they are to be multiplied. The equations predicting observed
variables from latent variables can be derived from this matrix expression by
progressing across each line and performing the operations. For x1, the three
terms in the first row of Λx matrix are multiplied by the elements in the ξ matrix
as follows, x1=
[ λ 11 0 0 ] [ ξ 1 ξ 2 ξ 3 ] = λ 11 ( ξ 1 ) + 0 ( ξ 2 ) + 0 ( ξ 3 ) = λ 11 ( ξ 1 )
.
Then, the error term δ1 is added, resulting in the equation given earlier:
x 1 = λ 11 ξ 1 + δ 1 /
.
In models with endogenous latent variables (e.g., Figure 2.4), the endogenous
latent variable equations have the same format but different notation, as
indicated earlier:
y1=λ7η1+ε1
(p.39)
y2=λ8η1+ε2
y3=λ9η1+ε3.
These equations can be expanded into matrix notation in the same way as the
exogenous latent variable equations.
Structural Model Equations. Figure 2.5 included two endogenous variables, one
of which (Parenting, η2) was predicted by an exogenous latent variable (Risk2,
ξ2), and one of which (Behavior, η1) was predicted by both the exogenous latent
variable (Risk1, ξ1), and the endogenous observed Parenting variable (η2). The
equations given earlier for these structural relationships were
Page 20 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
η 1 = β 12 η 2 + γ 11 ξ 1 + ζ 1
η 2 = γ 22 ξ 2 + ζ 2 .
The compact expression for these equations is
η=Bη+Γξ+ζ.
where B (capital β) is the matrix of β parameters between endogenous variables,
and Γ (capital γ) is the matrix of γ parameters between exogenous and
endogenous variables.
The matrix format corresponding to both the detailed and compact equations is
[ η 1 η 2 ] = [ 0 β 12 0 0 ] [ η 1 η 2 ] + [ γ 11 0 0 γ 22 ] [ ξ 1 ξ 2 ] + [ ξ 1 ξ 2 ]
.
If you carry out the operations, you obtain
η 1 = ( 0 ) η 1 + β 12 η 2 + γ 11 ξ 1 + ( 0 ) ξ 2 + ζ 1 ,
which reduces to the original equation for η1 above.
In summary, one important way that matrices are used in SEM is to convey the
elements and operations of equations that define SEM models.
Λ x = [ λ 11 0 0 λ 21 0 0 λ 31 0 0 0 λ 42 0 0 λ 52 0 0 λ 62 0 0 0 λ 73 0 0 λ 83 0 0 λ 93 ]
.
Page 21 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Although the rows and columns are not labeled, it is understood through the
subscripts that the rows correspond to observed x variables 1 through 9, and the
columns correspond to latent ξ variables 1, 2, and 3. We noted earlier that the
first λ subscript in factor equations referred to the (dependent) indicator
variable, and the second referred to the factor. The same rule applies for the Λ
matrix entries; the first subscript refers to the number of the indicator variable
or row, and the second refers to the number of the factor or column. In Figure
2.2, no observed variable (p.41) loaded on more than one factor. Consistent
with the figure, the Λx matrix above specifies that one factor loading is to be
estimated for each variable and the loadings for the other two factors are to be
fixed at 0. In confirmatory factor analysis, it is possible, however, to have
variables load on multiple factors. If, for example, observed variable 2 (x2)
loaded on factors 1 and 3 (ξ1, ξ3), and variable 6 (x6) loaded on factors 1 and 2
(ξ1, ξ2), the matrix for the model would be
Λ x = [ λ 11 0 0 λ 21 0 λ 23 λ 31 0 0 0 λ 42 0 0 λ 52 0 λ 61 λ 62 0 0 0 λ 73 0 0 λ 83 0 0 λ 93
]
.
A second matrix that is used in the analysis of a measurement model is the Φ
(capital phi) matrix, containing variances and covariances of the latent variables
(φ’s, phis). This matrix has one row and one column for each latent variable in a
model. The phi matrix for the model in Figure 2.2 with three correlated latent
variables, therefore, would look like the following:
Φ = [ φ 11 φ 21 φ 22 φ 31 φ 32 φ 33 ]
.
The phi matrix is symmetrical. Values above the diagonal are not included
because they are identical to those below the diagonal. The covariance of ξ1 and
ξ2, for example, is the same as the covariance between ξ2 and ξ1. As with a
covariance matrix of observed variables, the values on the diagonal are
variances. Again, the rows and columns are not labeled, but it is understood
through the subscripts that the values from left to right and from top to bottom
apply, respectively, to ξ1, ξ2, (p.42) and ξ3. If any pair of factors in a model do
not covary, a 0 would replace the corresponding off-diagonal φ element.
Page 22 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
The third matrix used in the analysis of measurement models is the Θδ (theta
delta) matrix, containing the error variances and covariances of the observed
indicators of exogenous variables (θ’s). The theta matrix has one row and one
column for each observed variable in the CFA model. The diagonal of the Θδ
matrix contains the variances of the error terms of observed variables, and the
off diagonals contain their covariances. Usually error terms are not correlated,
however, in CFA they are allowed to be, if there is theoretical justification. It is
considered reasonable, for example, to allow the errors of the same measure
administered at two different times to be correlated. Often, CFA models are
revised to include correlated errors to improve fit. This issue will be discussed in
more detail in Chapter 4. In the example of a Θ matrix following this paragraph,
most of the error covariances are fixed at 0, however, the matrix specifies that
the covariance between the error terms for variables 4 and 5 is expected to be
different from 0:
Θ δ = ( θ 11 0 θ 22 0 0 θ 33 0 0 0 θ 44 0 0 0 θ 54 θ 55 )
The estimates in the Λx,Φ, and Θδ matrices are used in SEM analyses to generate
an x by x matrix of estimated population variances and covariances (Σxx, sigma)
using the equation presented after this paragraph. The equation is based on a
sequence of algebraic proofs using matrix algebra and expectation theory, which
are beyond the scope of this book. Users interested in learning how the equation
was derived as a central expression in CFA are referred to Bollen (1989) and
Long (1983) for more information.
∑ xx = Λ x Φ Λ x ′ + θ δ
Long (1983) emphasizes the importance of this equation. It indicates how
estimated parameters of a confirmatory factor model can be (p.43)
manipulated into a new implied matrix of variances and covariances that can be
compared to the original matrix of observed variances and covariances. It is
important to remember that because the symbols are capital Greek letters, each
element of the equation represents a matrix (not just one number). In words, the
equation reads as follows:
Page 23 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Structural Model Matrices. So far, we have discussed the three matrices that are
used in the analysis of a confirmatory factor model. Up to five additional
matrices are used in the analyses of structural models. First, the factor loadings
of the indicators of dependent latent variables are contained in the Λy matrix,
which has the same properties as the previously discussed Λx matrix. The
variances of the error terms for the indicators of the dependent latent variables
are contained in a Θε (theta epsilon) matrix that has the same properties as the
Θδ (theta delta) measurement matrix. Note that the error variance of an
exogenous variable like Gender in Figure 2.4, which is assumed to be measured
without error, would be fixed to 0 and included in the Θδ matrix. The error
variance of Par10 in Figure 2.5, would also set to 0 if the endogenous latent
Parenting variable in that model was assumed to be measured without error by
Par10. If Par10 had a known reliability, its error variance could alternatively be
specified in the Θε matrix as 100% minus that reliability value.
[ γ 11 0 0 γ 22 ]
.
The γ11 parameter represents the path from Risk1 (ξ1) to Behavior (η1) that is
present in Figure 2.5. The 0 to its right represents the absence of a path from
Risk1 to Parenting—i.e., the fixing of the value of that path to 0. The 0 in the
second row represents the absence of a hypothesized path from Risk2 to
Behavior. The γ22 parameter represents the path from Risk2 to Parenting.
The fourth new matrix encountered in general structural models is the B (beta)
matrix, which contains the regression paths between pairs of endogenous (i.e.,
η) variables. This matrix has one row and one column for each endogenous
variable in a model. The B matrix for Figure 2.5 would look as follows:
[ 0 β 12 0 0 ]
.
Page 24 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
The final new matrix that is used in the estimation of structural models is the Ψ
(psi) matrix, which contains the variances and covariances of the structural
errors (i.e., ζ’s) in a model. Endogenous latent variables are not represented in
the Φ matrix of variances and covariances among ξ’s, and their variances are not
estimated. Instead, the variances of their associated error terms are estimated.
The values represent the amount of variance in the endogenous variables that is
unexplained by predictors in the model, and from these values the percent of
variance explained can be calculated. In Figure 2.5 there are two endogenous
structural variables (Behavior and Parenting). Each has a ζ term. The Ψ matrix
has one row (p.45) and one column for each endogenous variable. In most
cases, no correlation between ζ terms will be modeled, so off-diagonal elements
of the Ψ matrix will be 0. The diagonal of the matrix contains the variances of
the error associated with each endogenous variable. For Figure 2.5, this matrix
would look as follows:
( Ψ 11 0 0 Ψ 22 )
.
Some structural models with latent variables do not posit directional
relationships among endogenous latent variables; they may only have directional
relationships among exogenous and endogenous variables. In such cases, no B
matrix is needed.
We saw earlier that one equation, Σxx = ΛxΦΛx´+θδ, relates CFA model estimates
to the population covariance matrix of observed variables. For structural models
with Λy, Θε, Γ, B, and Ψ, matrices, the relationship is more complicated. A new
matrix based on four matrix blocks created by four equations relates estimates
of parameters in the eight SEM matrices to the new implied matrix of variance
and covariances. Notation for these equations varies across sources; we use
Bollen’s (1989) notations:
(∑YY=ΛY(Ι−B)−1(ΓΦΓ'+Ψ)[(Ι−B)−1]'ΛY'+Θε∑YX=ΛY(Ι−B)−
1(ΓΦΛ'X)∑XY=ΛXΦΓ'[(Ι−B)−1]Λ'Y∑XX=ΛXΦΛX'+Θδ)
Page 25 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Note that the lower right equation is the covariance equation used in CFA
models. Because CFA, or measurement-only, models have no η, β, γ, λ y or ε
values, and therefore no B, Γ, Λy, Θε, or Ψ matrices, only the third equation is
necessary to generate the comparison matrix in CFA models. For the derivation
of these equations based on matrix algebra and expectancy theory, see Bollen
(1989). Although it is not essential to know how these equations were derived, it
is important to understand that the equations permit the all-important linking of
estimated parameters to an implied matrix that can be compared to the original
covariance matrix of observed variables.
Page 26 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
(p.47) The null hypothesis in SEM is that the population covariance matrix
equals the matrix that is implied by the CFA or general structural model. The
equation for this null hypothesis in the population is
H0:∑=∑(θ)
The equation states simply that the population variance covariance matrix (Σ,
sigma) equals the implied matrix (Σ) that is based on esti-mated parameters
(contained in θ). (Note that θ here has a different meaning from the θ used to
designate the measurement error matrices.) Technically, this null hypothesis
invokes the inference of population values from sample statistics. Because the
population matrix is rarely available to researchers, however, the sample
covariance matrix (derived from the observed variables in our dataset) is
substituted in the equation (Bollen, 1989). Therefore, the equation for the null
hypothesis in the sample is
H0:S=∑(θˆ).
which states that the covariance matrix Σ(θˆ) reproduced based on parameter
estimates is not statistically different from the input matrix of observed
covariances for the sample (S). As described in Box 2.4, the SEM researcher
wants to accept the null hypothesis of no difference.
The null hypothesis in SEM analyses is that the input or analyzed matrix of
observed covariances is statistically the same as the implied or reproduced
matrix obtained by estimating parameters specified in the researcher’s
model. Unlike in an intervention study, for example, where the researcher
wants evidence that two group means are not the same, the SEM researcher
wants to accept the null hypothesis of no difference:
H0:S=∑(θˆ)
The difference between two matrices such as S and Σ(θˆ) can be presented
in a third matrix, the residual matrix, in which the elements indicate the
differences between corresponding elements in the input and implied
matrices.
Page 27 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
(p.48) The residual matrix is the matrix containing the differences between
corresponding elements in the analyzed and implied matrices. It is obtained by
subtracting each element of the implied matrix from its counterpart in the input
matrix. If the elements of a residual matrix are small and statistically
indistinguishable from 0, then the analyzed model fits the data well.
Before the input and implied matrices are compared to determine if the null
hypothesis can be accepted or must be rejected, the estimator attempts to
minimize the difference between the two matrices. An iterative estimation
process is used in which parameter estimates are obtained, tested, tweaked, and
tested again until no more reduction in the difference between the original and
implied matrices can be obtained. The determination that the two matrices are
as similar as possible is made through applying a “fitting” or “discrepancy”
function that quantifies the difference. The set of parameter estimates that
yields the smallest value for this discrepancy function becomes the final solution
for the model. When the smallest value is achieved, the estimation process has
converged on a solution in which the discrepancy function has been minimized.
The minimization value obtained is critical for assessing the hypothesis that the
input and implied matrices are statistically equivalent.
After the discrepancy function has been minimized, various tests are run to
determine just how similar the two matrices are, and whether the differences
are statistically significant. One test reported by all SEM (p.49) software is the
actual statistic obtained with the discrepancy function. Values obtained with the
fitting functions are χ2 (chi square) distributed, so they can be evaluated in
terms of statistical significance with regard to the number of degrees of freedom
(discussed in the following section on identification) of the model. A
nonsignificant χ2 value indicates that the null hypotheses can be retained—the
researcher’s model is consistent with the data. A statistically significant χ2 value
indicates that S and Σ(θˆ) are statistically different. However, due to limitations
of the χ2 statistic, there are now a large number of additional tests of fit that can
be used to support claims of good fit, even if the χ2 statistic is statistically
significant. Specific fit indices will be examined in Chapter 6.
Identification
Page 28 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Structural equation models are identified (generally) when there are more
covariances and variances in the input data matrix than there are parameters to
be estimated, and when each latent variable has a metric or measurement scale
(Kline, 2005). The amount of observed information available for an SEM model is
the number of unique elements in the covariance or correlation matrix being
analyzed. Model underidentification occurs when the number of parameters to
be estimated in a model (p.50) exceeds the number of unique pieces of input
data, or when there is too little information available for the estimation of any
one parameter. In SEM analysis, “just-identified” means that the number of
parameters to be estimated in a model is equal to the number of unique pieces
of input data. The difference between the number of unique matrix elements and
the number of parameters to be estimated is called the degrees of freedom of a
model.
Illustration of Identification
Count the number of observed variables in Figure 2.2. There are nine variables
represented with rectangles in the model, x1 through x9. A covariance matrix of
these variables would have 9 by 9 elements. However, the full matrix would have
fewer than 81 pieces of unique information because it contains redundant items.
For example, the covariance of x1 and x5 would be presented in the column
under x5; the covariance of x5 with x1, the same quantity, would be presented in
the column under x1. Instead of 81 pieces of information in a 9 by 9 covariance
matrix, there are p (p +1)/2, or 9(10)/2 = 45, unique pieces of information,
where p is the number of observed variables.
Page 29 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019
Structural Equation Modeling Concepts
Table 2.1 illustrates a covariance matrix of three variables (p = 3). The variances
of the three variables are along the diagonal and are nonredundant pieces of
information. The three covariances above and below the diagonal are redundant
—only one set should be counted. Using the formula p(p + 1)/2, there are 3(3 +
1)/2 = 6 unique pieces of information in the matrix—the number shown in the
shading in Table 2.1.
x1 x2 x3
(p.51) measurement model presented in Figure 2.2. There are nine observed
variables and nine factor loadings. One loading on each factor is typically fixed at 1 for
scaling the latent variable and identifying the latent variable. (More will be said about
this later.) Fixing one loading per factor reduces the number of parameters to be
estimated. With three factor loadings (one for each factor) fixed at 1, only six factor
loadings need to be estimated. There are three latent variables, so three variances
will be estimated. There are three interfactor covariances. There are nine error
variances, one for each observed variable. Therefore, 21 parameters need to be
estimated. The covariance matrix contains 45 unique elements. 45 – 21 equals a
positive number, so the model is identified. It has 45 – 21 degrees of freedom, or 24 df.
If the model had 0 degrees of freedom, that is, it is just-identified, only one solution is
possible and it will have perfect fit (the input and implied matrices will be equal).
Models with 0 degrees of freedom cannot be tested for their quality in comparison to
other models. Models with negative degrees of freedom (i.e., underidentified models)
will not run in most SEM programs—there are too few rules guiding the analysis, and
an infinite number of solutions may be possible.
Page 30 of 30
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (www.oxfordscholarship.com). (c) Copyright Oxford University Press, 2019. All
Rights Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a
monograph in OSO for personal use (for details see www.oxfordscholarship.com/page/privacy-policy). Subscriber: Utrecht
University Library; date: 27 March 2019