Upload
gifugifu
View
215
Download
0
Embed Size (px)
Citation preview
7/29/2019 TF14.SEM.
1/52
1
STRUCTURAL EQUATION MODELLING
MODERN METHODS
(Chapters 14, pages 681704, 714-735, 755 - 785)
Often applied psychologists are part of a research and development team
trying to design interventions to help people improve their quality of life.
Ideally, these interventions are based upon a theory that models aspects of
human behaviour within particular social contexts. The more precise the
theory and the more it is based upon empirical evidence, the more specific
components of an intervention can be designed based upon sound scientific
principles as well as upon the experience and intuition of dedicated
practitioners. Indeed, theory is useful both because it suggests ways to probe
the effectiveness of conventional practices and because it suggests innovative
practices that have not been considered in the past.
It bears repeating (stifle your yawns!), that modern theories must be complexin nature if they are to have relevance to individuals designing social
programs. As well, practitioners can only manipulate some of the pertinent
variables that influence program participants. Testing theory in this situation
is not addressed adequately by conventional hypothesis testing research
methods, even though these methods provide strong support for central parts
of the theory. More and more, researchers wish to test the plausibility of their
theoretical models as a whole; that is they wish to examine the interaction of
an entire set of variables in a field setting that includes important social
outcomes. The growing interest in structural equation modelling (SEM)
reflects this desire.
7/29/2019 TF14.SEM.
2/52
2
Whatever the SEM program selected by the researcher, the underlying
statistical procedure uses matrix algebra to solve a set of simultaneous
equations that are specified by a particular theory. These equations are
directly derived from a path diagram that shows the relationships among the
theoretical constructs (the structural model) and the relationships among the
theoretical constructs and the measures of these constructs (the measurement
model). Using these equations and initial start-up values, SEM programs
follow an iterative algorithm which converges on an optimal solution (it is the
different iterative procedures and their criteria for convergence that
distinguishes the various SEM programs). The extent to which this solution
can reproduce the variancecovariance matrix among the variables in the
data set becomes a test of the fit of the theoretical model (similar to the
factor analysis criterion that the factor solution should reproduce the
correlation matrix among the variables entered into the analysis).
Thus the results of a variety of goodness-of-fit indices are an important
outcome of any SEM program and it is possible to compare different theories
in terms of their relative goodness-of-fit. As well, the results of the analysis
can indicate whether adding certain paths will significantly improve the fit
between the model and the data. This is a valuable, albeit, post hoc procedure
that suggests specific modifications to the theory that warrant investigation
and replication in a subsequent study.
7/29/2019 TF14.SEM.
3/52
3
The degree to which a model fits the data is only one of the outcomes of
interest, however. Of equal interest are the values of path coefficients that
estimate the strength and direction of both direct and mediated relationships
among the variables in the model. Indeed, SEM can model mediating
psychological processes in a much more direct way than traditional
experimental designs.
Combining the results of experiments that establish a causal link from an
independent variable to a dependent variable with the results of SEM that
place this causal relationship within a theoretically specified network of
relationships provides far more compelling evidence in support of a theory
than either kind of evidence alone. The path coefficients are unbiased
estimates of population parameters and they can be tested to see if they are
significantly different from zero. As well, the relative value of standardized
path coefficients (range -1 through 0 to +1) indicate their importance in
predicting specific outcomes.
Another name for structural equation modelling is confirmatory factor
analysis. Researchers using this name are specifically concerned with
generating evidence that a psychological construct takes the form specified by
theory (often the construct is multidimensional; e.g., sexism has both a hostile
and a benevolent aspect which are related). This type of analysis is usually
used when the researcher is developing tests which measure the construct
adequately (with reliability, content validity, and construct validity).
7/29/2019 TF14.SEM.
4/52
4
To repeat, psychologists use structural equation modelling techniques to test
the adequacy of a theoretical model and to estimating the strength of causal
paths using path coefficients. This modelling technique also allows the
reliability of the measuring instruments to be estimated by specifying the
measurement model and a structural model and testing the viability of both
types of models simultaneously; a process that involves using a complex
iterative computer program to analyse a large data set.
In this course only a basic introduction to SEM as it is used to conduct a
confirmatory factor analysis or to assess the plausibility of a theoretical model
with recursive paths will be covered. As well, the exercises and assignment
will give you practice with one of the most widely used windows program,
EQS (installed on the Arts Lab computers in room 31 Arts). The Department
has a site licence for EQS so your supervisor can install it in her/his lab.
Note that if only one measured variable (manifest variable) is used to index
each underlying construct in a theoretical model, the SEM procedure conducts
a path analysisan analysis that only includes measured variables.
7/29/2019 TF14.SEM.
5/52
5
An Impor tant Note:In the past, researchers often assumed that theoretical
models are recursive. This is limiting now and will likely be even more
limiting in the future. People react to events with a sense of agency rather
than as pawns controlled by fate, and their actions count. The ability of
theory to model complex feedback loops and the ability of statistics to test the
plausibility of such theories is becoming increasingly important.
To use a clinical example, victims of violence often do not label themselves in
this way. Rather active coping attempts by these individuals allow them to
avoid applying the label of victim to themselves and to successfully adapt.
Understanding how some individuals are able to do this while others succumb
to this extremely stressful experience is essential if professional psychologists
are to help effectively.
Similarly, in social psychology, stereotyping is only of concern because this
invidious process is perpetuated through selective perceptions guided by
unreasonable expectations. Again, understanding the feedback loop that
results in group members confirming the negative expectations implied by
their groups stereotype is a complex problem that involves reciprocal causal
influences between stereotyper and stereotypee in an interpersonal interaction
that extends across time.
More positively, consider the development of friendship which, at its heart,
involves increasingly intimate and reciprocal interactions as the lives of two
people become inextricably bound together. As all these examples show,
reciprocal causality lies at the heart of many important psychological
processes and the requirements for statistical techniques to help model these
processes are in increasing demand.
7/29/2019 TF14.SEM.
6/52
6
CONVENTIONS USED IN SEM DIAGRAMS:
RECTANGLES = MEASURED OR MANIFEST VARIABLES
OVALS = CONSTRUCTS OR LATENT VARIABLES OR FACTORS
ARROW = CAUSAL PATH; DOUBLE ARROW = CORRELATION
EXOGENOUS VARIABLES = INDEPENDENT VARIABLES
ENDOGENOUS VARIABLES = DEPENDENT VARIABLES WHICH ARE
SOMETIMES INDEPENDENT VARIABLES AS WELL.
Canadian
Identity
Perceived
Discrimination
Cultural
Identity
Illegitimate
X Unstable
Intentions
(Protest)
Emotions
Em2
Em3
Em3
IDC3
IDC2
IDC1
PD3
PD2
PD1
ID1
ID2
ID3
exogenous
latent
variable
endogenous
latent
variable
E = Error
D = Disturbance
exogenous
manifest
variable
endogenous
manifest
variable
endogenous
manifest
variable
7/29/2019 TF14.SEM.
7/52
7
Assumptions Underlying Structural Equation Modelling
1. All variables must have linear relationships with each other.
2. Outliers must be identified and dealt with prior to the main analysis.
3. The variables in the analysis should be normally distribution(multivariate normality). This assumption is more crucial for modern
forms of SEM and so a preliminary analysis must involve examining each
variable for skewness and kurtosis. Transformations that create a normal
distribution for these variables are then applied. If a transformation does
not achieve a normal distribution, then an iterative estimation procedure
that is robust to violation of this assumption must be used.
4. Absence of multicollinearity is necessary as the computer executes matrixinversions in each iteration. Most SEM programs give the determinant of
the variancecovariance matrix as part of the output so that this
assumption can be examined.
7/29/2019 TF14.SEM.
8/52
8
5. Sometimes SEM programs have difficulty analysing a data set that containsvariables measured on scales that vary considerably in range and mean
value, a situation which results in covariances that are tremendously
different in size. Rescale some of the variables before running the analysis.
6. The possibility of a specification error haunts any researcher usingstructural equation modelling. The solution is to examine the residual
variancecovariance matrix. The residuals should be small and centred
around zero. Non-symmetrical residuals (some small and some large)
suggest that the model estimates some parameters well and others poorly.
One reason for this is that a causal path between variables in the model has
been mistakenly set to zero (the theory is wrong). If this is true, then post
hoc procedures can be used which suggest how the model can be
improved by adding paths. Then replication using another sample is
required. The other reason why residuals are large and non-symmetrical is
that the model is misspecified. There is no easy solution to this problem
but at least the analysis pushes the researcher to examine the theory more
critically.
7/29/2019 TF14.SEM.
9/52
9
7. Large sample sizes are required in order to run modern structuralequation modelling programs. Generally, the minimum sample size for all
SEM programs can be estimated by multiplying the number of
parameters that the program is estimating by ten. This means that usually
EQS requires a sample size of at least 200 research respondents and other
programs require more. However, experienced applied researchers with
messy data say that even that number may not be enough for the program
to convergence on a final solution. That is, with smaller sample sizes and,
therefore, more unstable estimates, the program simply may not be able to
find an optimal solution. Part of this problem can be caused by the default
start values used by the SEM program being very different from the actual
values of the parameters. Therefore, if estimates of these parameters can
be obtained from past research, they should be specified as the initial start
values in the analysis.
7/29/2019 TF14.SEM.
10/52
10
8. Structural Equation Modelling is based upon a mathematical procedurethat tests the ability of a theoretically derived model to reproduce the
variancecovariance matrix among measured variables in a data set.
The use of the variancecovariance matrix preserves the scale of the
original variables. Rescaling these variables by, for example, adding or
subtracting a constant does not change the results of the analysis.
However, rescaling variables through the use of sample statistics is more
problematic as it alters the value of the 2 statistic that is the basis fortesting the goodness-of-fit of the model. When variables are standardized,
the rescaling involves sample statistics because deviations from the meanare divided by the sample's standard deviation. Hence, the developer of
EQS, Peter Bentler, warns researchers not to use correlations whenever
possible. In some circumstances, a researcher has no choice because he or
she is doing a secondary analysis on a correlation matrix from a published
article. In this instance, the EQS program alerts users of the program to
the fact that the analysis may not be correct because a correlation matrix
has been used.
Of course standardized path coefficients are very useful because they
reflect the relative strength of different paths in the model. Thus, after the
analysis on the variancecovariance matrix has been done, the computer
calculates these standardized path coefficients from the unstandardized
path coefficients and their standard errors.
Perhaps in the future this limitation will be overcome, but right now it is
important for researchers to know that they should analyse the variance
covariance matrix, NOT the correlation matrix, whenever possible (the
default option in EQS).
7/29/2019 TF14.SEM.
11/52
11
Model Specification Using the Bentler-Weeks Estimation Method
The Bentler-Weeks model takes a regression approach to structural equation
modelling. However, the matrix equation specifies both measured variables
(also called manifest variables) and the latent variables (constructs or factors)
that are presumed to underlie responses to the measuring instruments. In this
model, both types of variables can be exogenous or endogenous.
Remember from chapter 5 that the matrix algebra equation for multiple
regression is:
y = x . b + e
In this equation there are k regression coefficients (in a k x 1 vector) which
need to be estimated and there are always enough equations to provide an
estimate of these parameters (as long as N > k). The reason why it is always
possible to estimate the regression coefficients is because it is assumed that 1)
the independent variables are measured without error, 2) the independent
variables have a direct causal influence on the dependent variable and no
other variables, and 3) the residuals associated with the dependent variable
are not correlated with the independent variables (no specification error). In
other words, these assumptions allow the research analyst to fix the values of
many parameters that could, theoretically, vary and it is the specification of
these parameters that allows the computer to estimate unique values for the
regression coefficients (the estimated parameters) using a least squares
solution.
7/29/2019 TF14.SEM.
12/52
12
In the same way, a more complex regression equation can be written
describing the relationships among the endogenous (dependent) and
exogenous (independent) variables specified by the structural (theoretical)
model and the measurement model. However, unlike multiple regression,
SEM asks the researcher to choose which parameters he or she will fix and
which to estimate. However, the researcher can not allow all possible
parameters to vary freely because this would always result in an under-
identified model (not enough degrees of freedom available to estimate the
parameters in the model). Thus the researcher must define both the
measurement model and the structural model in such a way that enough
parameters are fixed in value so that there are degrees of freedom available to
test the plausibility of the model in its entirety. That is, the model must be
over-identified.
Bearing this in mind, the fundamental Bentler-Weeks regression equation can
be expressed as:
= B . + . (q x 1) (q x q) (q x 1) (q x r) (r x 1)
where (eta) is a q x 1 vector of the q endogenous (dependent) variables; B(beta) is a q x q square matrix of path (regression) coefficients which are
estimates of the relationships among the endogenous variables; (gamma) isa q x r matrix of path coefficients which are estimates of the relationships
between the endogenous variables and the exogenous (independent) variables;
and (xi) is a r x 1 vector of the r exogenous variables.
7/29/2019 TF14.SEM.
13/52
13
Notice that this method involves solving q equations. That is, there is an
equation for each of the q endogenous variables and there are no equations
for the exogenous variables because their variability is explained by variables
outside the model. However, the r exogenous variables have variances and
covariances that need to be estimated. These variances and covariances are in
an r x r variancecovariance matrix called (phi).
Altogether, then, the parameters that need to be estimated are in the B, , and matrices and the path diagram is used to set some of these parameters tofixed values (usually 0 or 1) so that there are enough degrees of freedom
available to test the goodness of fit of the model to the data. Start values for
the parameters are then entered into the matrices. These start values can be
set by the computer or they can be estimated and entered by the researcher.
The computer then estimates the variancecovariance matrix among all the
measured variables using the criterion for convergence specified by the
researcher and compares it to the actual variancecovariance matrix. New
parameter estimates are calculated and entered as start values in the next
iteration. The computer stops the iterations when the estimate of the variance
covariance matrix cannot be improved.
Notice in this form of structural equation modelling 1) the B, , and matrices contain parameter estimates for both the measured and the latent
variables (factors) and 2) the and vectors are not estimated but deriveddirectly from the data set.
7/29/2019 TF14.SEM.
14/52
14
Using the Path Diagram to Derive Equations
Representing the Theory
Consider the path diagram shown below for the Skiing Satisfaction example
in TF (Fig 14.4, p. 692).
F2 = skisatF1 = loveski 1.0
V1 = yrsski
V2 = daysski
V5 = senseek*
V4 = foodsat
V3 = snowsat
*
E1* 1.0
*
1.0
*
D2*
1.0
*
1.0
E3*1.0
*
E4*1.0
*
1.0
*
1.0
*
1.0
*
1.0
1.0
*
1.0
In this diagram the stars (*) indicate a path coefficient or a variance that
needs to be estimated. The number of stars equals the total number of
parameters that need to be estimated (see later diagrams).
7/29/2019 TF14.SEM.
15/52
15
Consider the elements in this path diagram. Remember, measured variables
are represented by squares and latent variables (factors) by ovals.
Endogenous variables have causal paths leading to them, while exogenous
variables do not. Residuals of the endogenous measured variables are
included in the model as exogenous variables labelled E (for error), while
residuals of the endogenous latent variables (factors) are included in the
model as exogenous variables labelled D (for disturbancesthe error in
prediction). Notice that endogenous and exogenous variables can be either
manifest or latent. Also note that usually the path from the errors and
disturbances are fixed at 1, but their variance is estimated. This reflects the
preference of most researchers to estimate the residual error and to not be
concerned with estimating the path coefficient from unknown variables
outside the model.
By convention, this model fixes the variance of the exogenous latent variable,
love of skiing, at 1. This is the researcher's decision. Alternatively, he or she
could have set one of the path coefficients from this latent variable to one of its
indicators to 1, as is done for the path from ski trip satisfaction to snow
satisfaction. Essentially this latter convention gives the latent variable the
same scale as the chosen indicator and is advocated strongly by statisticians,
including Bentler. Allowing all the paths and the variance to vary freely is not
an option, however, as this will usually cause the model to become under-
identified.
7/29/2019 TF14.SEM.
16/52
16
How then are equations derived from the path diagram? First, it is important
to realize that the number of equations that you need to write is equal to the
number of endogenous variables. In this example, four measured variables
(number of years skied, number of days skied, snow satisfaction, and food
satisfaction) are endogenous and one latent variable, ski trip satisfaction, is
endogenous. Therefore, there are five endogenous variables represented in
five equations. Four of these equations represent the measurement model as
they link the latent factors to the measured variables. The fifth equation
represents the structural model specified by theory; in this case that the love
of skiing and sensation seeking are the causal determinants of ski trip
satisfaction.
Next you need to write the five equations in the same way as the matrix
equation:
= B . + .
The matrix equation for this example is shown below (see TF page 694), where
q = 5, r = 7, and * = parameter needs estimating:
V1 0 0 0 0 0 V1 0 * 1 0 0 0 0 V5
V2 0 0 0 0 0 V2 0 * 0 1 0 0 0 F1
V3 = 0 0 0 0 1 V3 + 0 0 0 0 1 0 0 E1
V4 0 0 0 0 * V4 0 0 0 0 0 1 0 E2F2 0 0 0 0 0 F2 * * 0 0 0 0 1 E3
E4
D2
7/29/2019 TF14.SEM.
17/52
17
Consider the first endogenous variable, Number of years skied (V1). In the
path diagram this variable is predicted by the latent factor, love of skiing, (F1)
and variables outside the model (E1). Thus, the equation for this variable is:
V1 = * F1 + 1 . E1
In its full form, however, it would be written:
V1 = (0 . V1 + 0 . V2 + 0 . V3 + 0 . V4 + 0 . F2) + (0 . V5 + * . F1 + 1 . E1 + 0 . E2 + 0 . E3 + 0 . E4 + 0 . D2)
Remember when two matrices are multiplied each element of the first row in
the first matrix is multiplied by the corresponding element in the first column
of the second matrix and then added, and so on. Thus, the full form of this
equation is equivalent to the first line of the matrix equation shown above.
7/29/2019 TF14.SEM.
18/52
18
In the same way, the other equations for the endogenous variables are
equivalent to the corresponding lines of the matrix equation specified by the
Bentler-Weeks SEM model. This means that writing these equations is
equivalent to writing the matrix equation for this SEM procedure and writing
these equations is easily done using the path diagram. The five equations
implied by this particular path diagram are:
V1 = * F1 + E1
V2 = * F1 + E2
V3 = 1 F2 + E3
V4 = * F2 + E4
F2 = * F1 + * V5 + D2
Note that in these equations the path coefficients from E1, E2, E3, E4, and D2
are set to 1.
In addition, the model specifies the variances and covariances among the
exogenous variables that need to be estimated This means that the variances
for V5, E1, E2, E3, E4, and D2 need to be estimated, that the variance of F1
is set to 1, and that all the covariances are set to 0, specifying the variance
covariance matrix, (see page 695 which shows this diagonal matrix).
Together this information is sufficient for the computer to execute the
iterative estimation procedure (section 14.4.4 shows how the computer goes
through one of these iterations). Specifically, start values replace the stars (*)
in these matrices. Then the SEM program estimates the variance
covariance matrix among the measured variables.
7/29/2019 TF14.SEM.
19/52
19
Methods of Achieving Convergence
So far we have discussed how, in general, the computer converges on a best
estimate of the variance covariance matrix for the measured variables.
However, there are several different criterion for achieving convergence. The
most common of these is the Maximum Likelihood (ML) method followed by
the Generalized Least Squares (GLS) method. Essentially, the Maximum
Likelihood method converges on an estimated variancecovariance matrix
which maximizes the probability that the difference between the estimated
and the samples variance - covariance matrices occurred by chance.
In contrast, the Generalized Least Squares method converges on an estimated
variancecovariance matrix that minimizes the sum of the squared
differences between the elements in the estimated and the samples variance -
covariance matrix.
Mathematically, both of these criteria for convergence involve minimizing a
mathematical function through successive approximations. Both methods are
good if the variables are distributed normally and the sample size is adequate.
Tabachnick and Fidell suggest using an estimation method called the scaled
Maximum Likelihood method if non-normality can not be corrected. EQS will
do this if required and gives a corrected Chi squared called the Satorra-
Bentler scaled 2 (TF, p. 713).As well, adjustments to the standard errors of
the path coefficients are calculated so as to correct their statistical
significance.
(I will discuss using dichotomous variables later.)
7/29/2019 TF14.SEM.
20/52
20
Testing the Adequacy of the Theoretical Model
(including Goodness-of-Fit Indices)
The Basic Chi Squared Test for Goodness-of-Fit
Once the SEM program has converged on a solution, a 2 statistic is calculatedthat tests how well the estimated variance - covariance matrix fits the actual
variance - covariance matrix among the measured variables (its value is taken
from the value of the mathematical function that is minimized to achieve
convergence). The degrees of freedom for this statistic are equal to the amount
of unique information in the sample variancecovariance matrix minus the
number of parameters that need to be estimated. If there are p measured
variables, then the total degrees of freedom, p*
= p ( p + 1) / 2.
In the ski trip satisfaction study, for example, there are five measured
variables so the total number of degrees of freedom are 5 x 6 / 2 = 15. Given
that the path diagram shows that the researcher wants to estimate five path
coefficients and six variances, 2 is tested with 15 - 5 - 6 = 4 degrees offreedom.
7/29/2019 TF14.SEM.
21/52
21
Because the desired result is for a good fit between the estimated and the
sample variancecovariance matrices, the researcher hopes for a non-
significant 2 . However, 2 values are dependent upon sample size such thateven very small differences are significant when the sample size is large. The
result is that a number of Goodness-of-fit indices have been developed to
correct for this problem and many of them are part of the output in the EQS
program.
Testing the Significance of the Path Coefficients
In the Bentler-Weeks SEM procedure, the unstandardized path coefficients
are normally distributed so that when they are divided by their standard
error a Z score is obtained. It is these Z scores that provided a test of
whether the path coefficient is significantly different from zero.
As the unstandardized path coefficients are on a different scale from one
another, researchers often report the standardized path coefficients that vary
from -1 to 0 to +1 and which indicate the relative strengths of the causal paths
in the model in the same way as a standardized regression coefficient. Notice
that the standardized path coefficients from the latent variables to their
measured counterparts are factor loadings. Indeed, if the measurement
model specifies the entire model then the SEM procedure has executed a
confirmatory factor analysis as will be illustrated shortly in an example.
7/29/2019 TF14.SEM.
22/52
22
Goodness-ofFit Indices
Given the problems associated with using the lack of significance of a 2statistic to indicate a models goodness-of-fit to the data, many other indices
have been developed. Unfortunately, there is no consensus on which of these
indices are the best ones to use, so SEM programs avoid this problem by
outputting most of these indices. Usually, they all show that the model is or is
not a good fit, so it is a matter of preference which one the researcher reports.
Hu & Bentler (1999) argue, however, that researchers should usually report
one residual based fit index and one comparative fit index. The following
section discusses some of the more commonly used goodness-of-fit indices of
both types.
7/29/2019 TF14.SEM.
23/52
23
Comparative Fit Indices
In SEM the theoretical model can be compared to a just identified model that
contains all possible causal paths. A just identified model can not be tested,
but the matrix equation will give an exact and unique mathematical solution.
If one more parameter needs to be estimated, then the mathematical solution
is indeterminate (there are not enough equations to yield a solution). More
commonly the theoretical model is compared to a model in which all the
variables are independent of one another. Here all path coefficients are set to
zero and only the values in the variancecovariance matrix for the exogenous
variables are estimated.
These two extremes illustrate the fact that models vary along a continuum
from a model in which all the variables are independent of one another (only
the variances of the exogenous variables are estimated) to the just identified
model which can be specified but which can not be tested because there are no
degrees of freedom left. The theoretical models goodness-of-fit is estimated by
comparing it with one or the other of these extremes.
A simple and often used comparative goodness-of-fit index is the Bentler
Bonett Normal Fit Index (NFI) defined as:
NFI = 2indep - 2model / 2indep
This index varies from 0 to 1 and values greater than 0.9 indicate a good fit.
7/29/2019 TF14.SEM.
24/52
24
However, for relatively small samples ( < 200) the NFI underestimates the fit
of a model. Thus, it has been replaced by the Comparative Fit Index (CFI)
which uses a noncentrality parameter, , which is an index of modelmisspecification; the larger the value of the greater the misspecification,with = 0 indicating that the estimated model is perfect.
The Comparative Fit Index is defined by the following equation:
CFI = 1 - (model / indep )where = (2 - df)
This index also varies from 0 to 1 and is a better estimate of goodness-
of-fit for smaller samples. Models are a good fit if CFI > .95.
Another popular comparative fit index is the Root Mean Square Error
of Approximation (RMSEA) which, compares the model to a just identified
model. This statistic is defined as
RMSEA = ( Fo / dfmodel )
And Fo = (2model - dfmodel) / N , with Fo set to zero if its value is negative.
Fo = 0 indicates a perfect fit, so small values of RMSEA are desired. A good
fitting model is indicated if RMSEA < 0.06. However, like the NFI, this
statistic tends to reject models that fit well when the sample size is small.
7/29/2019 TF14.SEM.
25/52
25
Residual-Based Fit Indices
A popular index of goodness-of-fit that has intuitive appeal is the Root
Mean Square Residual (RMR) index. This statistic is based upon the average
of the squared differences between each element of the sample variance -
covariance matrix and the corresponding element of the estimated variance
covariance matrix.
RMR = ( 2 ij ( (sj - j)2
/ p ( p - 1) )
,
where p is the number of measured variables and sj and j are thecorresponding variances and covariances from the two matrices.
Models that fit well have small RMR values, but these values are dependent
upon the scale of the original measured variables in the model. Therefore, a
standardized Root Mean Square Residual index (sRMR) has been developed.
The values for sRMR range from 0 to 1 with small values indicating a good fit
(small residuals). sRMR < .08 indicates that the model is a good fit.
(In the social psychology literature the CFI and the sRMR are fit indices that
are often reported.)
7/29/2019 TF14.SEM.
26/52
26
Model Identification
Can the Theoretical Model be Tested?
In order to test any model, it has to be over-identified. This means that there
is a unique solution to the mathematical procedure which results in estimates
for all the parameters that are allowed to vary freely in the model, and that
there is at least one degree of freedom available to test this model using chi
squared. If there are p measured variables, then the total number of degrees
of freedom is, p*
= p ( p + 1) / 2. The number of parameters that need to be
estimated must, therefore, be less than this number.
However, if the model describes relationships among latent variables as well
as the relationships among these factors and measured variables, then the
SEM program may still not be able to converge on a solution. This is because
both the structural model and the measurement model must be over-identified
in order to test the entire models goodness-of-fit.
Once p*
has been calculated, the next step is to establish whether the
measurement model is likely to be over-identified. If there is one latent
factor in the model, there needs to be three variables measuring this construct
and their errors must be uncorrelated. If there are two or more latent factors,
the same conditions apply provided that each set of three measured variables
only load on one factor and that the factors are allowed to covary. Sometimes
two indicators per factor is sufficient under these conditions provided that
none of the variances and covariances among the factors are zero.
7/29/2019 TF14.SEM.
27/52
27
Note that, in order for the factor to have meaning, one of the measured
variables is used to scale the latent variable by setting the path coefficient to 1
(the variance of the latent variable is the same as the measured variable).
This is called a marker variable. Failure to set the scale of a factor is one
common error which results in identification problems.
Looking at whether the structural model is over-identified is the next step in
this process. Provided there is only one latent variable or that the latent
variables are recursive and their disturbances do not correlate, this part of the
model is likely to be identified.
Notice that the phrase likely to be over-identified is used when discussing
both the measurement model and the structural model. This is because the
guidelines just reviewed do not guarantee that a particular model can be
tested. Establishing this with certainty is complex, so perhaps the best
strategy is to apply these guidelines and then run the analysis. The EQS
program signals when this problem has arisen by indicating that some
parameters are linearly dependent on other parameters.
(NOTE: Dun et al. (1993) suggest using p*
= p (p + 1) / 2 and then running the
analysis to see if problems arise. If they do, more parameters can be given
fixed values to deal with the problem.)
7/29/2019 TF14.SEM.
28/52
28
Using Parcels of Measured Variables When the
Sample Size is Small
Sometimes applied researchers are forced to do structural equation modeling
with a relatively small N (less than 200). In this instance, the researcher must
balance the need to include enough measured variables to adequately specify
the measurement model with the need to restrict the number of parameters
estimated by the model as a whole. The solution is to create a small number of
parcels made up by averaging the responses to several of the original
measured variables (e.g., questionnaire items) with the minimum of three
parcels per latent variablethe number usually required for any SEM
program to run properly.
Parcels of items (measured variables) are constructed for each latent
variable using the following item-to-construct balance method:
Consider parceling measured variables (items) measuring a construct into
three parcels. First a factor analysis is done on all the items measuring the
construct. Then the three items with the highest factor loadings are used to
anchor the three parcels. The three items with the next highest loadings are
added to the anchors in reverse order, and so on. Together these three parcels
are used as the manifest variables in the SEM analysis rather than the original
items (see Little, Cunningham, Shahar, & Widaman, 2002). This reduces the
number of parameters that the SEM program needs to estimate for the
measurement model. However, this should only be done if N is small (usually
< 200). Otherwise it is better to use the original measured variables in the
analysis as multiple indicators of the construct.
7/29/2019 TF14.SEM.
29/52
29
EXAMPLE OF A SIMPLE CONFIRMATORY FACTOR ANALYSIS
USING THE EQS PROGRAM
Confirmatory Factor Analysis is a factor analytic technique that is design to
test theory that specifies an underlying structure to a construct. Instead of
discovering the underlying factor structure in a post hoc fashion through the
exploratory factor analysis techniques covered earlier in this course, the
function of this type of factor analysis is to confirm that the theorized factor
structure underlying a construct is plausible. When SEM is used purely for
confirmatory factor analysis, the theory that defines a construct in a certain
way is tested, but the relationship of that construct to other constructs is not
explored. This means that SEM is used to test a measurement model.
Self-concept is a complex multi-dimensional psychological construct that
psychologists have been interested in since our discipline began. In the
example, a two factor theory of academic self-concept is specified which
suggests that it is comprised of two underlying and interrelated components
reflecting different aspects of the self: English self-concept (ESC) and maths
self-concept (MSC). Each of these components can be measured in several
ways and that the responses to these measures are caused by these two factors
in the manner specified by the path diagram on the next page. SEM tests this
overall model as well as the specified causal paths which represent hypotheses
derived from this academic self-concept theory. The study is a secondary
analysis of published data summarizing the responses of 996 adolescents to a
self-concept test battery. The authors of the study have provided the
variance-covariance matrix and so EQS is used to analyse the data in this
matrix.
7/29/2019 TF14.SEM.
30/52
30
The Path Diagram Specified by Theory
E12*
V3
V9
V10
ESC*
E3*
E9*
E10*
V4
V11
V12
MSC*
E4*
E11*
1.0
1.0
*1.0
*
1.0
1.0
1.0
*
1.0
*
1.0
*
7/29/2019 TF14.SEM.
31/52
31
The EQS Syntax File ( *.EQS)
/TITLE
Self-concept: Confirmatory Factor Analysis Example
The ti tle statement can be several l ines long and can contain explanatory notes
on the decisions that r esul ted in the syntax being used.
/SPECIFICATIONS
VARIABLES= 12; CASES= 996;
DATAFILE='c:\data\Eqs\807 2004\SEM notes.CFA example.byrne.ess';
MATRIX=COR;
ANALYSIS = COV;
METHOD=ML;
The specif ications commands give the computer details on the number of
vari ables in the data set (VAR), the sample size (CASE),the location of the data
matri x (DATAF I LE -- a * .ESS is a EQS data fi le), the type of data matr ix being
analysed (MATRIX = COR or COV); the basis for the analysis (ANALYSIS =
COR or COV), with the default being the variancecovariance matri x, and the
iterative estimation procedure (METHOD). Al l subcommands are separated by
semi -colons (a general EQS syntax rule except for data matrices).
The data is in the form of a correlation matri x with standard deviationsV1 V2 V3 V4 V5 V6 V7 V8 V9 V10 V11 V12
1.0000 0.0710 0.2890 0.1700 0.8630 0.4786 0.2522 0.2160 0.1560 0.1280 0.1770 0.1350
0.0710 1.0000 0.2450 0.2530 0.2660 0.3060 0.2619 0.7675 0.2420 0.3070 0.2475 0.3040
0.2890 0.2450 1.0000 0.0120 0.2270 0.2990 0.2389 0.3430 0.7050 0.8543 0.0660 0.0270
0.1700 0.2530 0.0120 1.0000 0.2000 0.2250 0.3460 0.3472 0.0140 0.0690 0.8640 0.8280
0.8630 0.2660 0.2270 0.2000 1.0000 0.8310 0.7100 0.2160 0.1900 0.1310 0.2700 0.1880
0.4786 0.3060 0.2990 0.2250 0.8310 1.0000 0.2537 0.2830 0.2100 0.1740 0.2570 0.1870
0.2522 0.2619 0.2389 0.3460 0.7100 0.2537 1.0000 0.2545 0.1440 0.1396 0.2426 0.0367
0.2160 0.7675 0.3430 0.3472 0.2160 0.2830 0.2545 1.0000 0.2690 0.2900 0.2489 0.3057
0.1560 0.2420 0.7050 0.0140 0.1900 0.2100 0.1440 0.2690 1.0000 0.7627 0.1420 0.0280
0.1280 0.3070 0.8543 0.0690 0.1310 0.1740 0.1396 0.2900 0.7627 1.0000 0.0960 0.14600.1770 0.2475 0.0660 0.8640 0.2700 0.2570 0.2426 0.2489 0.1420 0.0960 1.0000 0.8060
0.1350 0.3040 0.0270 0.8280 0.1880 0.1870 0.0367 0.3057 0.0280 0.1460 0.8060 1.0000
The second last row is the standard deviations of the vari ables
14.1000 12.3000 10.0000 16.1000 9.3000 14.9000 9.4000 15.3000 11.3000 15.7000 11.5000 12.4000
The last row contains the means which, in thi s example, are set to zero.
7/29/2019 TF14.SEM.
32/52
32
EQS Syntax (Continued)
/LABELSV3 = ESC1; V4 = MSC1;
V9 = ESC2; V10 = ESC3; V11 = MSC2; V12 = MSC3;
F1 = ESC; F2 = MSC;
These commands give more meaningful variable labels than those used by the
computer. The syntax also helps you wr ite the equations which need to use the
computer labels. Dont forget the semi -colons.
/EQUATIONSV3 = F1 + E3;
V9 = *F1 + E9;
V10 = *F1 + E10;
V4 = F2 + E4;
V11 = *F2 + E11;
V12 = *F2 + E12;
These equations specify the model that is being tested. Each equation is
separated by a semi-colon. I f you run EQS using a path diagram (EQS
diagrammer f unction), these equations wil l be generated fr om the path diagramautomatically.
/VARI ANCES
F1 TO F2 = *;
E3 TO E4 = *; E9 TO E12 = *;
/COVARIANCES
F1 TO F2 = *;
/END
These commands specify the matr ix of var iances and covariances. Notice thatthe covari ation among the error terms are not specif ied implying that they are set
to zero (the defaul t). The /END statement tell s the computer to begin the
analysis.
7/29/2019 TF14.SEM.
33/52
33
The EQS Output File (*.OUT)
The f irst page of the output repeats the syntax that was just presented.
TITLE: Self-concept: Confirmatory Factor Analysis Example
COVARIANCE MATRIX TO BE ANALYZED: 6 VARIABLES (SELECTED
FROM 12 VARIABLES) BASED ON 996 CASES.
This line reminds the researcher that the model contains 6 of the original 12
variables in the data set. Successive runs of EQS could specify dif ferent subsets
of data for dif ferent analyses.
Then the output gives the enti re variance-covar iance matr ix among the variables
used in the analysis.
ESC1 MSC1 ESC2 ESC3 MSC2
V 3 V 4 V 9 V 10 V 11
ESC1 V 3 100.000
MSC1 V 4 1.932 259.210
ESC2 V 9 79.665 2.547 127.690
ESC3 V 10 134.125 17.441 135.311 246.490
MSC2 V 11 7.590 159.970 18.453 17.333 132.250
MSC3 V 12 3.348 165.302 3.923 28.423 114.936
MSC3
V 12
MSC3 V 12 153.760
I f a correlation matr ix was analysed, the computer reminds the researcher
because this is not a good idea.
CORRELATI ON MATRIX TO BE ANALYZED:
7/29/2019 TF14.SEM.
34/52
34
BENTLER-WEEKS STRUCTURAL REPRESENTATION:
NUMBER OF DEPENDENT VARIABLES = 6
DEPENDENT V'S : 3 4 9 10 11 12
Here the endogenous (dependent) variables in the model are identified.
NUMBER OF INDEPENDENT VARIABLES = 8
INDEPENDENT F'S : 1 2
INDEPENDENT E'S : 3 4 9 10 11 12
Here the exogenous (independent) variables in the model are identified.
NUMBER OF FREE PARAMETERS = 13
This is the number of parameters the researcher is estimating in thi s analysis
(stars in the equations plus the stars in the vari ancecovar iance matrix , or
equivalently the number of stars on the path diagram).
The total number of degrees of freedom in this data set is calculated by the
formula p*
= p (p + 1) / 2 where p is the number of measured variables. In
this example, there are 6 measured variables, so the total number of degrees
of freedom are 6 x 7 / 2 = 21. As the researcher wishes to estimate 13
parameters, the degrees of freedom remaining that can be used to test the
goodness-of-fit of the model is 21 - 13 = 8.
DETERMINANT OF INPUT MATRIX IS 0.99223E+11.
Clearl y mul ticoll ineari ty is not a problem in this data set.
AVERAGE ABSOLUTE STANDARDIZED RESIDUALS = 0.0182
AVERAGE OFF-DIAGONAL STANDARDIZED RESIDUALS = 0.0255
The computer then pri nts out the residual variancecovariance matri x and the
standardized residual var iancecovar iance matri x (not shown). Following each
matri x is an average of al l the residuals and all the off diagonal residuals. These
averages should be small if the model f its the data well . The average of the off
diagonal residuals is given because smal l residual covariance values are more
crucial for the model to be a good fi t.
7/29/2019 TF14.SEM.
35/52
35
LARGEST STANDARDIZED RESIDUALS:
V11, V9 V12, V10 V4, V3 V 9, V4 V12, V3
0.076 0.069 -0.064 -0.054 -0.044
The computer then pr ints out the 20 largest residual values (thi s is an extract) sothat the researcher knows which relationshi ps are not modelled very well . For
example, the fi rst and largest residual in th is table shows that the model does
not explain a relationship between an index of Engl ish self concept (V9) and
math self -concept (V11) as well as other relationships in the samples correlation
matri x. Whether the researcher wil l use this information or not depends on the
overall f it of the model and the size of these residual covariances (look at
standardized residual s > .10).
DISTRIBUTION OF STANDARDIZED RESIDUALS
----------------------------------------
! !
20- -
! !
! !
! !
! ! RANGE FREQ PERCENT
15- -
! ! 1 -0.5 - -- 0 .00%
! ! 2 -0.4 - -0.5 0 .00%
! * ! 3 -0.3 - -0.4 0 .00%! * ! 4 -0.2 - -0.3 0 .00%
10- * - 5 -0.1 - -0.2 0 .00%
! * * ! 6 0.0 - -0.1 12 57.14%
! * * ! 7 0.1 - 0.0 9 42.86%
! * * ! 8 0.2 - 0.1 0 .00%
! * * ! 9 0.3 - 0.2 0 .00%
5- * * - A 0.4 - 0.3 0 .00%
! * * ! B 0.5 - 0.4 0 .00%
! * * ! C ++ - 0.5 0 .00%
! * * ! -------------------------------
! * * ! TOTAL 21 100.00%----------------------------------------
1 2 3 4 5 6 7 8 9 A B C EACH "*" REPRESENTS 1 RESIDUALS
Th is hi stogram shows that the residuals are centred on zero (100% are between
0.1 and 0.1) and are symmetr ical. This information indicates that the model
does not contain a ser ious specif ication error.
7/29/2019 TF14.SEM.
36/52
36
GOODNESS OF FIT SUMMARY
INDEPENDENCE MODEL CHI-SQUARE = 5093.587 ON 15
DEGREES OF FREEDOM
This first chi square test should be signif icant as it test the hypothesis that thevari ables are independent of one another (one of the standards of comparison
used by some of the comparative goodness-of -f it indices).
CHI-SQUARE = 266.589 BASED ON 8 DEGREES OF FREEDOM
PROBABILITY IS LESS THAN 0.000001
This is the basic chi square value that tests the goodness-of-f i t of the model.
Because the sample size is large (N = 996), the fact that th is statistic i s
signif icant does not mean that the model i s a poor fi t.
BENTLER-BONETT NORMED FIT INDEX= 0.948
BENTLER-BONETT NONNORMED FIT INDEX= 0.905
COMPARATIVE FIT INDEX (CFI) = 0.949
These fit i ndices and parti cularly the CFI suggest that the model i s qui te a good
f it as they are all around the 0.9. The CFI should be greater than 0.95 for the
model to be a good fi t.
ITERATIVE SUMMARY
PARAMETER
ITERATION ABS CHANGE ALPHA FUNCTION
1 69.848500 1.00000 .48517
2 5.597344 1.00000 .28123
3 .887007 1.00000 .26795
4 .069267 1.00000 .26793
5 .012263 1.00000 .26793
6 .000785 1.00000 .26793
This output shows how the function specif ied by the estimation method
converges on a minimum value. Noti ce that the chi square testing the goodness-
of-f it of the model is equal to the minimum function value mul tipl ied by (N-1):
0.26793 x 995 = 266.589 (with in rounding error).
7/29/2019 TF14.SEM.
37/52
37
The computer then wr ites the equations for the endogenous variables with the
estimated parameters.
MEASUREMENT EQUATIONS WITH STANDARD ERRORS AND TEST
STATISTICS
ESC1 =V3 = 1.000 F1 + 1.000 E3
MSC1 =V4 = 1.000 F2 + 1.000 E4
These fir st set of equations are the ones in whi ch the paths of one of the
indicators (a marker) for the four factors was fi xed at a value of 1. This gives
the underlying factor the same scale as the marker variable.
ESC2 =V9 = 1.010*F1 + 1.000 E9
.031
32.704@
This equation shows that the unstandardized path coeff icient between the
Engli sh self -concept factor and the specif ic measure of Engl ish self -concept
(ESC2) is 1.010. The standard error for th is statistic is 0.031 and the Z score is
1.010 / .031 = 32.704. As this Z score is greater than 1.96, it is signi fi cant as
indicated by @.
ESC3 =V10 = 1.705*F1 + 1.000 E10.040
42.362@
MSC2 =V11 = .696*F2 + 1.000 E11
.014
49.339@
MSC3 =V12 = .720*F2 + 1.000 E12
.016
44.579@
Sometimes, one estimated parameter is linearly dependent upon the others. The
computer wil l run the analysis, but wil l print out an error message concerning
this linear dependency. Do NOT tr ust the output if thi s message appears. A less
extr eme form of li near dependence is shown by a parameter estimate having a
very smal l standard error (relative to standard errors found in past research).
Consider eliminating thi s var iable and rerunning the analysis.
7/29/2019 TF14.SEM.
38/52
38
The computer then pr ints the estimated variances and covariances of the
exogenous variables with standard errors and Z scores. The estimates for the
measur ed var iables shou ld be inspected to see if they seem reasonable given your
knowledge of these variances and covari ances from past research. Sometimes
the estimation procedure produces an odd solution that does not conform to the
f indings fr om past research. I n this instance, it is probably wise to give more
credence to the resul ts of past research until the resul ts of the analysis are
replicated.
VARIANCES OF INDEPENDENT VARIABLES (EXTRACT)
----------------------------------
I F1 - ESC 78.690*I
I 4.539 I
I 17.338@I
I E3 - ESC1 21.310*I I
1.524 I I
13.980@I I
As an extr eme, the SEM program can estimate negative variances for measured
vari ables. This is fl agged by the computer with an error message saying that
these variances can not be estimated. I n this instance, the anal ysis is seriously
f lawed and the researcher wil l need to reassess whether the use of SEM to
analyse the data set is warranted. As well, the estimated variances of the latent
exogenous factors can be negative. The computer wil l not allow this to happen
and wil l constrain the variance to zero (or a lower bound estimate that is
posi tive). I f you see thi s error message, ser iously question the validity of the
results.
mailto:17.338@Imailto:17.338@I7/29/2019 TF14.SEM.
39/52
39
STANDARDIZED SOLUTION: R-SQUARED
ESC1 =V3 = .887 F1 + .462 E3 .787
MSC1 =V4 = .941 F2 + .338 E4 .886ESC2 = V9 = .793*F1 + .610 E9 .628
ESC3 = V10 = .963*F1 + .269 E10 .928MSC2 =V11 = .918*F2 + .397 E11 .842
MSC3 =V12 = .879*F2 + .476 E12 .773
These standardized path coeff icients from the measur ed vari ables to the latent
factors are the ones that are most usuall y wri tten onto the path diagram in
publ ished reports (they are calculated fr om the unstandardized path coeff icientsafter the analysis is completed). Note that thi s table can not be calculated if some
of the estimated var iances for the exogenous variables are negative.
I n this example, the standardized path coeff icients are the factor loadings of the
measured var iables on the latent factors. The squared mul tiple correlati ons (the
square of the path coeff icient) are estimates of the proportion of the var iance of
the measur ed var iables which is shared with the underlying factor. Thi s is a
commonali ty estimate for the variable on the factor and, equivalentl y, an
estimate of i ts reli abili ty. For example, the reli abili ty of the Math Self -Concept
Scale (V12) is .773. This number indicates the proportion of variance in the
measured variable that measures the under lying Maths Self -concept construct
(F2). (This number is NOT a good reli abil ity estimate if the error terms for the
measur ed variables are correlated. )
Notice that in thi s table the standardized path coeff icients for the marker
vari ables (and the error terms) that were fixed in the equations now have a value
dif ferent f rom 1 due to the standardization procedure. I f you need their statistical
signif icance, set another measur ed var iable as the marker and rerun the
analysis. This wil l give you the same value for the path coeff icients because thetwo solu tions are equivalent. The standardized variances are not printed as the
computer sets all variances equal to 1.
The corr elations among the latent factors is given in the last table of the output.
I n th is case it is 0.091 indicating that English and Math Academic Self -Concept
are relatively independent of one another.
7/29/2019 TF14.SEM.
40/52
40
Additional Syntax
The following syntax allows you to output additional goodness-of-fit indices.
FIT = ALL;
For this example, some of the output from this command is:
ROOT MEAN SQUARED RESIDUAL (RMR) = 5.133
STANDARDIZED RMR = 0.032
ROOT MEAN SQ. ERROR OF APP.(RMSEA)= 0.180
90% CONFIDENCE INTERVAL OF RMSEA ( 0.162, 0.199)
Post Hoc Adjustment of the Theoretical Model:
Addition and Subtraction of Parameters
In the above example, the value of the goodness-of-fit indices suggest that the
model is a good fit because they meet the established criteria. However, in
some analyses the solution looks quite good but could be improved (the fit
indices approach the criteria for a good fit, but do not meet these criteria). In
this situation, the researcher can conduct post hoc tests which suggest which
parameters should be estimated rather than fixed and which parameters can
be removed (set to zero). Then the modified model can and should be tested
on a new sample. These post hoc procedures capitalize on chance and so
Tabachnick and Fidell suggest only adding or subtracting a few paths one at a
time. As well they advocate using a conservative significance level (p < .01)
for selecting modifications to the parameters specified by theory.
7/29/2019 TF14.SEM.
41/52
41
Lagrange Multiplier Test
This post hoc procedure indicates which parameters could be added to the
model (estimated) to improve its goodness-of-fit based upon the current
sample. Both univariate and multivariate tests are conducted but the
multivariate test is the more important one as it identifies the parameters that
could be added into the model in a stepwise fashion similar to forward
selection in multiple regression. To run this procedure use the syntax:
/LMTEST.
Using this syntax in the context of the current example yields the following
output:
MULTIVARIATE LAGRANGE MULTIPLIER TEST BY
SIMULTANEOUS PROCESS IN STAGE 1
PARAMETER SETS (SUBMATRICES) ACTIVE AT THIS STAGE ARE:
PVV PFV PFF PDD GVV GVF GFV GFF BVF BFF
This line indicates the type of parameter matr ices that were active at this stage of
the analysis. The f irst letter i ndicates the matri x containing the suggested
parameter (P = ; G =; and B = B) and the remaining letters indicate thetype of var iables involved (V = measur ed variables, F = factors, E = errors, and
D = disturbances).
7/29/2019 TF14.SEM.
42/52
42
CUMULATIVE MULTIVARIATE STATISTICS UNIVARIATE INCREMENT
---------------------------------- -----------------------------
HANCOCK'S
SEQUENTIAL
STEP PARAMETER CHI-SQUARE D.F. PROB. CHI-SQUARE PROB. D.F. PROB.
---- ----------- ---------- ---- ----- ---------- ----- ---- ---
1 V4,F1 14.601 1 .000 14.601 .000 8 .067
2 V3,F2 25.807 2 .000 11.206 .001 7 .130
This part of the output indicates the possible changes along with a2test which,if signif icant, indicates that the model wil l be improved. For example the
analysis suggests adding a path fr om F1 (the Engl ish self -concept factor to V4
(supposedly a measure of M ath self -concept). Whil e statisticall y this makes
sense, theoretically it may not. I ndeed, whether the researcher actuall y makes
this change and recomputes the model depends on a thoughtful analysis of the
theoretical impli cations. One or two theoreticall y meaningful changes may well
improve the model suf f iciently to become a good fi t and th is strategy is usual ly
better than implementing all the changes without regard for theory.
Whenever changes are made, it is important to rerun the analysis on the
modif ied model so as to check on its goodness-of-f it and to examine the impact
of the changes on all the parameter estimates. I ndeed, if the modif ications
resul t i n parameter estimates that are not consistent with past research, the
researcher may decide that these modif ications are not worth making at all .
Af ter all , this is a post hoc procedure relying on purely statistical cr iteria. There
is no guarantee that the changes it suggests are changes that improveunderstand of the phenomena under study.
The Wald Test
This post hoc procedure is used to delete parameters (set them to zero)
and so make the model more restrictive. It is usually done after parameters
have been added using the Legrange test as adding paths changes the
parameter estimates. The following syntax is used to activate this procedure:
/WTEST
Because parameters are being set to zero, results of the 2 test should be non-significant. The output is similar to the Lagrange test. In this instance, no
paths were dropped when this test was conducted.
7/29/2019 TF14.SEM.
43/52
43
Comparing Nested Models: The Chi Square Difference Test
Sometimes different theories (or different versions of the same theories)
specify two models such that one model is nested inside the other. In this
instance the two models can be directly compared to see if the larger model
(the one with more paths) significantly improves the goodness-of-fit (or,
equivalently, if the added restrictions significantly reduces the goodness-of-
fit). This comparison is achieved by subtracting the 2 values for the twomodels. The result is another 2 statistic with degrees of freedom equal to thedifference in the degrees of freedom for the two models. This procedure
requires the estimation of two models, but its advantage is that it is theory-
based and provides evidence that directly bears upon the relative merits of the
two theories. In my view, this is a whole lot better than fixing a model in a
post hoc fashion using the Legrange Multiplier and/or the Wald tests.
7/29/2019 TF14.SEM.
44/52
44
EXAMPLE OF TESTING A CAUSAL MODEL
USING THE EQS PROGRAM
This path diagram specif ies a simple theoretical model of job satisfaction (an
endogenous latent variable) which was tested on 122 employees in an i ndustr ial
sales force. The exogenous latent variables in this model are achievement
moti vation and self-esteem.
SAT1
SAT2
ACH1
ACH2
SE1
SE2
F1
ACH*
SE*
1.0
E4*
*E5*
1.0
1.0E6*1.0
*
E7*1.0
*
E3*1.0
1.0
E2*1.0*
D1*
1.0
*
*
1.01.0
*
1.0
1.0
*
1.0
*
1.0
1.01.0*
1.0
*
*
7/29/2019 TF14.SEM.
45/52
45
The syntax used to test this theoretical model is derived directly from the
path diagram:
/TITLE
PERFORMANCE AND JOB SATISFACTION IN AN INDUSTRIAL SALES
FORCE
/SPECIFICATIONS
CASES = 122; VARIABLES = 8; MATRIX=CORRELATION;
ANALYSIS=COVARIANCE; METHOD = ML;
The number of parameters being estimated is 15, so the sample size (CASES =
122) is a li ttle smal l (i t should be at least 15 x 10 = 150).
The ANALYSIS command specif ies that the variance-covariance matri x should
be analysed. This matr ix is created fr om the corr elation matri x and the standard
deviations of the variables contained in the command statement /STA below.
/LABELS
V2 = SAT1; V3 = SAT2; V4 = ACH1;
V5 = ACH2; V6 = SE1; V7 = SE2;
F1 = JOBSAT; F2 = ACH; F3 = SE;
/EQUATIONS
V2 = F1 + E2;V3 = *F1 + E3;
V4 = F2 + E4;
V5 = *F2 + E5;
V6 = F3 + E6;
V7 = *F3 + E7;
F1 = *F2 + *F3 + D1;
The start values can be specif ied (f rom past research) for some or all of the
parameters in these equations (they are given as numbers to the left of the stars.
e.g., 0.5*F2). Specifying these start values is more crucial if you have a smal l
sample size.
7/29/2019 TF14.SEM.
46/52
46
/VARIANCES
F2 TO F3 = *;
E2 TO E7 = *;D1 = *;
/COVARIANCES
F2,F3 = *;
These sets of statements specif y the var iances and covariances for the
exogenous variables. Covari ances are set to zero by defaul t, so it is onl y
necessary to state that the covar iance between F2 and F3 needs to be estimated.
/MATRIX
1.000
.418 1.000
.394 .627 1.000
.129 .202 .266 1.000
.189 .284 .208 .365 1.000
.544 .281 .324 .201 .161 1.000
.507 .225 .314 .172 .174 .546 1.000
-.357 -.156 -.038 -.199 -.277 -.294 -.174 1.000
/STANDARD DEVIATIONS
2.09 3.43 2.81 1.95 2.08 2.16 2.06 3.65
This is the way a matrix of correlations with standard deviations (so as to
create a covariance matrix for the computer to analyze) is specified in the
command file.
/END
This statement ends the commands and tells the computer to begin the
analysis.
7/29/2019 TF14.SEM.
47/52
47
The EQS Output
The output starts out by repeating the syntax. Then the variancecovariancematr ix among the measur ed variables is given, followed by:
BENTLER-WEEKS STRUCTURAL REPRESENTATION:
NUMBER OF DEPENDENT VARIABLES = 7
DEPENDENT V'S : 2 3 4 5 6 7
DEPENDENT F'S : 1
NUMBER OF INDEPENDENT VARIABLES = 9
INDEPENDENT F'S : 2 3
INDEPENDENT E'S : 2 3 4 5 6 7
INDEPENDENT D'S : 1
NUMBER OF FREE PARAMETERS = 15
NUMBER OF FIXED NONZERO PARAMETERS = 10
The number of degrees of f reedom are p (p + 1) / 2 = 6 x 7 / 2 = 21. Therefore,
the degrees of freedom that can be used to test the goodness-of-f i t of the model is21 - 15 = 6.
DETERMINANT OF INPUT MATRIX IS 0.82146E+04
This shows that there is no problem with multi coll ineari ty.
The computer then pri nts out the residual variancecovariance matri x and the
standardized residual matr ix . The summary of the values in the standardized
matr ix shows that the residuals are smal l i ndicating that the model f i ts the data
well:
AVERAGE ABSOLUTE STANDARDIZED RESIDUALS = 0.0113
AVERAGE OFF-DIAGONAL STANDARDIZED RESIDUALS = 0.0159
7/29/2019 TF14.SEM.
48/52
48
The hi stogram of the residuals show that they are smal l and centred around zero
(over 97% li e in the range 0.1 to0.1).
GOODNESS OF FIT SUMMARY
CHI-SQUARE = 3.915 BASED ON 6 DEGREES OF FREEDOM
PROBABILITY VALUE FOR THE CHI-SQUARE STATISTIC IS 0.68813
COMPARATIVE FIT INDEX (CFI) = 1.000
This part of the output shows that the model i s a very good fi t and the chi square
is not signi f icant. The CFI also shows this and is the statistic to report given the
small sample size.
The computer then wr ites the equations for the endogenous variables with the
parameter estimates and their signif icance.
7/29/2019 TF14.SEM.
49/52
49
MEASUREMENT EQUATIONS WITH STANDARD ERRORS AND
TEST STATISTICS
SAT1=V2 = 1.000 F1 + 1.000 E2
SAT2=V3 = .929*F1 + 1.000 E3
.189
4.931@
ACH1=V4 = 1.000 F2 + 1.000 E4
ACH2=V5 = 1.006*F2 + 1.000 E5
.361
2.784@
SE1=V6 = 1.000 F3 + 1.000 E6
SE2=V7 = .879*F3 + 1.000 E7
.222
3.965
JOBSAT=F1 = .733*F2 + .547*F3 + 1.000 D1
.376 .234
1.949@ 2.335@
The output also estimates the variance and covariances of the exogenous
vari ables, including the corr elation between F2 and F3 as 0.396.
STANDARDIZED SOLUTION: R-SQUARED
SAT1= V2 = .743 F1 + .669 E2 .553
SAT2= V3 = .843*F1 + .537 E3 .711
ACH1=V4 = .622 F2 + .783 E4 .387
ACH2=V5 = .587*F2 + .810 E5 .344
SE1= V6 = .770 F3 + .638 E6 .593
SE2= V7 = .709*F3 + .705 E7 .503
JOBSAT= F1 = .349*F2 + .357*F3 + .808 D1 .347
7/29/2019 TF14.SEM.
50/52
50
Fur ther Comments on Structural Equation M odell ing with EQS
The Equivalency of Solutions Using Different Marker Variables
If there is more than one latent factor in your path diagram, then changing
the variable used as the marker variable for a factor results in an equivalent
solution (the size of the path coefficients are the same). However, if one
manifest variable is a much better index of the construct than the others, it is
best to use it as the marker variable as the solution tends to be more stable.
Setting the variance of exogenous factors to 1.0 (standardizing the variance)
rather than one of the paths to a marker variable is another option if you do
not want to specify a variable as a marker variable. This also results in an
equivalent solution.
The bottom line is that your choice of marker variable or your choice to set
the variance of a exogenous factor to 1.0 and not specify a marker variable
depends upon what you are interested in theoretically. For example, if you
want to scale a factor to a well known and highly reliable instrument, then you
should make this measured variable your marker variable. If all measuresare equivalent (and perhaps of unknown reliabilitye.g., face valid measures
of the construct) and you are not concerned with scaling a latent exogenous
factor, but rather want to know how all the measures load on this factor, then
set the variance of this exogenous factor to 1.0.
7/29/2019 TF14.SEM.
51/52
51
Including Dichotomous Variables in the Theoretical Model
Truly nominal variables define different groups of respondents withoutranking them (see TF, section 14.5.7, p. 730). Provided the sample size is large
enough, the SEM strategy is to test the theoretical model within each group
separately. If the model is supported in both samples, then your results
generalize across samples. For example, the model is supported for both men
and women; within white, immigrant, and Aboriginal samples; etc
A more advanced form of SEM not covered in this course or TF, starts by
testing the model within each sample and then doing a multiple group analysis
in order, for example, to test the invariance of the factorial structure of a
theoretical construct across groups. Simply put, the analysis constrains
certain parameters within each group to be equal (e.g., the size of the path
from a measured variable to a latent construct is the same for both men and
women) and examines whether the goodness-of-fit is still as good as the
goodness-of-fit of the model when these constraints are not applied.
If the sample size is small, you can include a nominal variable in the path
diagram as one or more dummy variables. Clearly these dummy variablesare not normally distributed so you have to use a robust estimation procedure.
If you are using ordinal data which reflects a underlying continuous variable
(e.g., age: 1 = young, 2 = middle aged, 3 = old), you must estimate the size of
the correlations that would have been obtained if you had actually measured
the continuous variable directly. These estimates are called polychoric
correlations (between two ordinal variables) or polyserial correlations
(between an ordinal and an interval variable) and they form the basis of the
analysis. This goes far beyond the scope of this course and is only briefly
mentioned in TF (section 14.5.6, p. 734).
7/29/2019 TF14.SEM.
52/52
52
Multivariate Kurtosis
If you run EQS using raw data, the program will print out Mardias
coefficient which indicates multivariate kurtosis (use p < .001 to determine if
kurtosis is a problem). The normalized estimate for this coefficient allows you
to see if the variables in the data set are normally distributed or not. For
large sample sizes, the values of the normalized coefficient correspond to Z
values and so large values indicate some non-normality is present due to
kurtosis.
The computer also prints out the 5 cases that have the largest normalized
estimate (and which contribute the most to the overall value of Mardias
coefficient). If one or two of these cases have much larger normalized
estimate values than the others, consider dropping these cases and re-running
the analysis. However, it is better not to drop these cases immediately, but to
check the variables for univariate and multivariate outliers and to adjust the
variables that have a non-normal distribution using transformations. This is
usually done with the SPSS program. EQS reads SPSS data files and converts
them into ***.ess files with ease.