Confirmatory Factor Analysis Rick DeShon
2008
OverviewWho am I?
Course purpose and focusUnderstand principles of modern measurementFocus on modelling and inference
Topic CoverageCourse website:
www.msu.edu/course/psy/818/deshon
Open time for general questions
Measurement!Most constructs in the social sciences are fuzzy, ill-defined, and unobservable
How do you measure something you don't fully understand and can't see?
Success in the physical sciences gives us hope and confidence
Ex: TemperatureEx: pH
Descriptive StatisticsSummation notation:
Mean:
Variance:
Covariance:
1
n
ii
x=∑
;x x
XN n
μ = =∑ ∑
( )( ) ( )( )2 2;1
i i i ix x x x x xs
N nμ μ
σ− − − −
= =−
∑ ∑
( )( ) ( )( );
1i x i y i i
xy xy
x y x x y ys
N nμ μ
σ− − − −
= =−
∑ ∑
Example - Univariateperson x x.dev x.dev.sq
1 18.32 8.04 64.612 5.55 -4.74 22.463 9.55 -0.74 0.554 9.02 -1.26 1.595 11.55 1.26 1.606 1.02 -9.27 85.867 13.68 3.40 11.558 22.57 12.28 150.929 -1.00 -11.28 127.32
10 12.59 2.30 5.31
Sum 102.85 0.00 471.76
Mean= Variance= Standard Deviation=10.29 52.42 7.24
Histogram of x
x
Fre
quen
cy
-5 0 5 10 15 20 25
0.0
0.5
1.0
1.5
2.0
2.5
3.0
Example - Bivariateperson x y x.dev y.dev xy.CrossProd
1 18.32 32.63 8.04 17.82 143.252 5.55 -4.15 -4.74 -18.97 89.883 9.55 22.26 -0.74 7.45 -5.514 9.02 -0.98 -1.26 -15.80 19.925 11.55 12.74 1.26 -2.07 -2.626 1.02 23.38 -9.27 8.57 -79.377 13.68 23.13 3.40 8.32 28.278 22.57 28.43 12.28 13.62 167.339 -1.00 -13.47 -11.28 -28.28 319.13
10 12.59 24.16 2.30 9.35 21.53
Covariance = 77.98
0 5 10 15 20
-10
01
02
03
0
x
y
Variance – Covariance Matrix
2
2x xy
yx y
σ σσ σ⎡ ⎤
Σ = ⎢ ⎥⎣ ⎦
2
2
52.42 77.9877.98 244.66
x xy
yx y
s sS
s s⎡ ⎤ ⎡ ⎤
= =⎢ ⎥ ⎢ ⎥⎣ ⎦⎣ ⎦
Covariance is unbounded and hard to interpretWhat is a big covariance?
Correlation transforms the covariance so that it ranges from -1 to +1
For our example...
Correlation
2 2
( )( ) ( )
xyxy
x y
Cov xyrVar x Var y
σ
σ σ= =
77.98 .68952.42* 244.66xyr = = 1.0 .69
.69 1.0r
⎡ ⎤= ⎢ ⎥⎣ ⎦
Correlation matrix
Simple RegressionSimple linear regression is the foundation of confirmatory factor analysis (CFA)So let's review it ...
0 5 10 15 20
-10
010
2030
x
y
Recall from algebra that the equation for a line is:
Slope = mrise/run or change in y for a unit change in x
Y-intercept = bValue of y when x=0
Simple regression applies this notion to bivariate data that are not perfectly related
Simple Regression
( )y m x b= +
What is the line that “best” fits or characterizes the relationship between x and y?
Population:
Sample:
Simple Regression
0 5 10 15 20
-10
010
2030
x
y
0 1i i iy xβ β ε= + +
0 1i i iy b b x ε= + +
Person 2: e = (-11.92)
Intercept and slope are constant across individualsy,x,and e are random variables that are likely different for each individual
Simple Regression
person x y1 18.32 32.632 5.55 -4.153 9.55 22.264 9.02 -0.985 11.55 12.746 1.02 23.387 13.68 23.138 22.57 28.439 -1.00 -13.47
10 12.59 24.16
1; 32.63 .49 1.48(18.32) 5.862; 4.15 .49 1.48(5.55) 11.92
10; 24.16 .49 1.48(12.59) 5.92
ii
i
= = − + += − = − + −
= = − + +
The variable means in most social science research are arbitraryTherefore, we may focus on analyzing data that are in deviation form (i.e., centered)
Simple Regression
person x.dev y.dev1 8.04 17.822 -4.74 -18.973 -0.74 7.454 -1.26 -15.85 1.26 -2.076 -9.27 8.577 3.4 8.328 12.28 13.629 -11.28 -28.28
10 2.3 9.35 -10 -5 0 5 10
-30
-20
-10
010
x.dev
y.de
v
When variables are in deviation form:The mean of each variable = 0.0
The intercept in the regression = 0.0
The regression slope is unchanged
The fit of the model is unchanged
The regression model reduces to:
Simple Regression
1i i iy xβ ε= +
To be consistent with CFA, we'll use slightly different notation to represent the same linear regression model
This relationship or equation may be represented graphically as:
Simple Regression
i i iy xλ ε= +
x y
e
λ
2eσ
lambdaλ =
EstimationOrdinary Least Squares (OLS)
AssumesExpected value of residuals = 0Homoskedasticity (constant variance) of residualsResiduals are independent of each other and xNormality of residuals
If met, OLS yields an unbiased, minimum variance estimate of the population slope
Simple Regression
2( )( )
xy y
x x
Cov xy rVar x
σ σλ
σ σ= = =
77.98ˆ 1.4852.42
λ = =
EstimationNotice that all the information we need to estimate the regression slope is available in the variance-covariance matrix
covariance of x and yvariance of x
Simple Regression
2
( )( )
xy
x
Cov xyVar x
σλ
σ= =
Hypothesis TestingGiven the magnitude of the estimate( ), is it reasonable to conclude that the population parameter ( ) is different from zero?
Standard error of the regression slope
Simple Regression
λ̂
λ
( )2
ˆ 2 1x Nε
λ
σσσ
=−
2 mean square error
root mean square errorε
ε
σ
σ
=
=
Hypothesis TestingThe slope estimate and standard error may be used to form a confidence interval or a significance test
Confidence interval for the regression slope
Significance test for the regression slope
Simple Regression
ˆˆ t
λλ σ±
ˆ
ˆt
λ
λσ
=
Example...High school and beyond dataset
200 observations from a sample of high school studentsdemographic information
gender (female)socio-economic status (ses)ethnic background (race)
scores on standardized testsreading (read)writing (write)mathematics (math)social studies (socst).
Simple Regression
Example...Let's examine the relationship between Science achievement scores and SES (numeric)
Simple Regression
. .. 0.525 2.028. 2.028 98.028
SES c SCI cS SES c
SCI c
⎡ ⎤⎢ ⎥= ⎢ ⎥⎢ ⎥⎣ ⎦
. .. 1.0 0.283. 0.283 1.0
SES c SCI cr SES c
SCI c
⎡ ⎤⎢ ⎥= ⎢ ⎥⎢ ⎥⎣ ⎦ -1.0 -0.5 0.0 0.5 1.0
-20
-10
010
20
SES.c
SC
I.c
Example...Let's regress Science achievement scores onto SES (numeric)
95% CI:
Simple Regression
( ). .i i iSCI c SES cλ ε= +
ˆ
ˆ 3.8670.9299.497
λ
ε
λσσ
===
3.867 4.16*0.929
t = =
( ).05 1.978t α= =
3.867 1.978*0.929 2.029 5.704to± =
Example...
Simple Regression
-1.0 -0.5 0.0 0.5 1.0
-20
-10
010
20
SES.c
SC
I.c
Not multiple regression!Multiple regression model
Multivariate simple regression model
Multivariate Simple Regression
1 1 2 2i i ii p p iy x x xλ λ λ ε= + + + +…
1 2 3, , , , ; 1...i i i ip j i iy y y y x j pλ ε= + =…
Multivariate Simple Regression
x y3
y2
y1
y4
yp
e1
e2
e3
e4
ep
λ1
λ2
λ3
λ4
λp
1
2eσ
2
2eσ
3
2eσ
4
2eσ
2peσ
Represents a system of linear equationsA separate regression equation for each outcome variable Moving toward a matrix representation...
Multivariate Simple Regression
1 1 11
2 1 21
1
0 00 0
0 0p p p
y xy x
y x
λ ελ ε
λ ε
⎡ ⎤ ⎡ ⎤ ⎡ ⎤⎡ ⎤⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥= +⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥
⎣ ⎦⎣ ⎦ ⎣ ⎦ ⎣ ⎦
Y X ε= Λ +or
1 ; 1...j j jy x j pλ ε= + =Capital Lambda
EstimationIf the errors are independent you can estimate each equation separately using OLS!!
If the errors are contemporaneously correlated, then this is a Seemingly Unrelated Regressions (SUR) estimation problem
If you estimate the system of equation simultaneously, you can impose equality constraints across the equations
More on this later...
Multivariate Simple Regression
Let's examine the HSB dataset from this perspectiveWhat's the effect of SES on the academic achievement test scores
All in deviation form (centered)SES.cSCI.cREAD.cWRITE.cMATH.c SOC.c
Multivariate Simple Regression
Covariance Matrix (S)
Correlation Matrix (r)
Multivariate Simple Regression
SES.c SCI.c READ.c WRITE.c MATH.c SOC.cSES.c 0.52 2.03 2.18 1.42 1.85 2.58SCI.c 2.03 98.03 63.97 53.53 58.50 49.44READ.c 2.18 63.97 105.12 58.00 63.61 68.41WRITE.c 1.42 53.53 58.00 89.84 54.83 61.54MATH.c 1.85 58.50 63.61 54.83 87.77 54.76SOC.c 2.58 49.44 68.41 61.54 54.76 115.26
SES.c SCI.c READ.c WRITE.c MATH.c SOC.cSES.c 1.00 0.28 0.29 0.21 0.27 0.33SCI.c 0.28 1.00 0.63 0.57 0.63 0.47READ.c 0.29 0.63 1.00 0.60 0.66 0.62WRITE.c 0.21 0.57 0.60 1.00 0.62 0.60MATH.c 0.27 0.63 0.66 0.62 1.00 0.54SOC.c 0.33 0.47 0.62 0.60 0.54 1.00
Run the multivariate regression analysisglm SCI.c READ.c WRITE.c MATH.c SOC.c with SES.Results...
Multivariate Simple Regression
3.867*4.152*2.715*3.524*4.919*
⎡ ⎤⎢ ⎥⎢ ⎥
Λ = ⎢ ⎥⎢ ⎥⎢ ⎥⎢ ⎥⎣ ⎦
9.4979.8029.2729.014
10.130
εσ
⎡ ⎤⎢ ⎥⎢ ⎥
= ⎢ ⎥⎢ ⎥⎢ ⎥⎢ ⎥⎣ ⎦
Notice that these estimates may be computed directly from the variance covariance matrix!
Multivariate Simple Regression
SES.c SCI.c READ.c WRITE.c MATH.c SOC.cSES.c 0.52 2.03 2.18 1.42 1.85 2.58SCI.c 2.03 98.03 63.97 53.53 58.50 49.44READ.c 2.18 63.97 105.12 58.00 63.61 68.41WRITE.c 1.42 53.53 58.00 89.84 54.83 61.54MATH.c 1.85 58.50 63.61 54.83 87.77 54.76SOC.c 2.58 49.44 68.41 61.54 54.76 115.26
1( . , . ) 2.03 3.90
( . ) .52Cov SES c SCI c
Var SCI cλ = = = Slightly different
due to roundingerror
1Variance of y not due to xεσ =
( )29.497 98.028 3.867 *.5246= −
Finally!!CFA is a multivariate regression model where you don't directly observe the exogenous (X) variable(s)Measurement theory
Can't directly observe many interesting phenomenae.g., intelligence, personality, temperament, beliefse.g., pH, temperature, nuclear particles
So, identify observable phenomena that are thought to be caused by the underlying variable that is latent or unobservable
i.e., temperature vs thermometer
Confirmatory Factor Analysis
CFA model is constructed in advancespecifies the number of (latent) factorsspecifies the pattern of loadings on the factors
factor loadings can be constrained to be zero (or any other value)
specifies the pattern of unique variancesmeasurement errors may be correlated
covariances among latent factors can be estimated or constrainedmultiple group analysis is possible
Can TEST whether the constraints are consistent with the data
Confirmatory Factor Analysis
Unidimensional CFA
y1
y2
y3
y4
ξ
δ1
x y3
y2
y1
y4
yp
e1
e2
e3
e4
ep
λ1λ
2
λ3
λ4
λp
σε1
σε2
σε3
σε4
σεp
Compare withMultivariate Regression
λ1
λ2
λ3
λ4
δ2
δ3
δ4
ksi or xi
1
2δσ 2
2δσ 3
2δσ 4
2δσ
2ξσ
Mathematical representation
y = (q × 1) vector of indicator/manifest variableq is the number of manifest variables
ξ = (p × 1) vector of latent variables or factorsksi or xip is the number of latent variables
Λ= (q × p) matrix of factor loadings (lambda)
δ = (q × 1) vector of errors of measurement (delta)
Unidimensional CFA
y ξ δ= Λ +
Mathematical representation
Unidimensional CFA
y ξ δ= Λ +
[ ]1 1 1
2 2 21
3 3 3
4 4 4
yyyy
λ δλ δ
ξλ δλ δ
⎡ ⎤ ⎡ ⎤ ⎡ ⎤⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥ ⎢ ⎥ ⎢ ⎥= +⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎣ ⎦ ⎣ ⎦ ⎣ ⎦
Mathematical representation2 additional variance-covariance matrices are needed
Unidimensional CFA
[ ]1 1 1
2 2 21
3 3 3
4 4 4
yyyy
λ δλ δ
ξλ δλ δ
⎡ ⎤ ⎡ ⎤ ⎡ ⎤⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥ ⎢ ⎥ ⎢ ⎥= +⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎣ ⎦ ⎣ ⎦ ⎣ ⎦
2ξσ⎡ ⎤Φ = ⎣ ⎦
Phi (pxp)
1
2
3
4
2
2
2
2
0 0 0
0 0 0
0 0 0
0 0 0
δ
δ
δ
δ
σ
σ
σ
σ
⎡ ⎤⎢ ⎥⎢ ⎥
Θ = ⎢ ⎥⎢ ⎥⎢ ⎥⎣ ⎦
Theta-delta (qxq)
Mathematical representationThe specified model results in an implied or predicted variance-covariance matrix, Σθ
Unidimensional CFA
θΣ = ΛΦΛ +Θ
21 1 11
22 1 1 2 1 22
23 1 1 3 2 1 3 1 33
24 1 1 4 2 1 4 3 1 4 1 44
λ φ θλ λφ λ φ θλ λφ λ λ φ λ φ θλ λφ λ λ φ λ λ φ λ φ θ
⎡ ⎤+⎢ ⎥+⎢ ⎥ΛΦΛ +Θ =⎢ ⎥+⎢ ⎥
+⎣ ⎦
If the unidimensional model is reasonable, then:
The covariance of scores on any two indicators is:
The product of the respective regression slopes (loadings) and the variance of the latent variable
The variance of the jth indicator is:
The sum of the squared loading and the error variance of the jth indicator
Unidimensional CFA
1
2ij i j ξσ λ λ σ=
1
2 2 2jjj j ξ δσ λ σ σ= +
Can also reverse this processIf we know the indicator variances and covariances, we can determine the model parameters that generated them
The regression slopes (loadings) may be determined from the covariance of any 3 indicators
This means that we must have at least 3 indicators to estimate the regression slopes!!
3 indicators = just identified single factor model
Unidimensional CFA
1
jk jlj
kl ξ
σ σλ σ σ=
Can also reverse this processThe error variances (aka uniquenesses) may be determined via subtraction
Unidimensional CFA
1
2 2 21j jjδ ξσ σ λ σ= −
Problem!Different covariance triplets will likely yield different estimates of the regression slopesWhich one is right...if any?
Solution – Maximum Likelihood estimationML is a very general estimation technique that determines the parameter estimates that maximize the fit between the implied covariance matrix and the actual covariance matrix
Minimizes the difference
Unidimensional CFA
Goal: Specify a model with an estimated covariance matrix that is not significantly different from the sample covariance matrix
Parameters are estimated simultaneously yielding an estimated covariance matrix, sigma_thetaS is the sample covariance matrix
The residual matrix is found by subtracting sigma_theta from S
Process iterates until changes in estimates do not improve model fit
Unidimensional CFA
Iterations continue until the likelihood is minimized
this is called convergenceWhen the model converges you get estimates of all of the parameters and a residual matrixχ2 is based in the function minimum (from ML estimation) when the model converges
The minimum value is multiplied by N – 1
Unidimensional CFA
Identification (for single factor models)Key issue-> do the parameters in the model result in a unique covariance matrix?
If different parameter sets yield identical outcomes, then there is no way to determine which set of parameters is superior
Def: The set of parameters θ={Λ,Φ,Θ} is not identified if there exists θ1≠θ2 such that Σ(θ1)= Σ(θ2)
This becomes a bigger issue when we get to multidimensional CFA
Unidimensional CFA
Identification (for single factor models)Necessary conditions:
T-rulep(p + 1)/2 - (# of estimated parameters)>0
The latent variable scale must be provided using one of two common constraints
Variance of the latent variable is fixed to 1Yields standardized loadings
Loading of one indicator is fixed to 1Yields unstandardized loadings
Unidimensional CFA
The unidimensional CFA model is a theoryYou believe that the observed variables are manifestations or indicators of a common latent variable that is not directly observableThe indicators may also reflect other sources of variance that are not common across the set of indicators = (uniqueness) or error variance (δ
i)
The variances and covariances among the observed variables are due to a single underlying latent (unobservable) variable
This is a very strong model of reality!
Unidimensional CFA
Example: High school and beyond dataVariance-covariance matrix for the 5 achievement tests
Correlation matrix for the 5 achievement tests
What latent variable might underly scores on these 5 achievement tests?
Unidimensional CFA
SCI.c READ.c WRITE.c MATH.c SOC.cSCI.c 98.03 63.97 53.53 58.50 49.44READ.c 63.97 105.12 58.00 63.61 68.41WRITE.c 53.53 58.00 89.84 54.83 61.54MATH.c 58.50 63.61 54.83 87.77 54.76SOC.c 49.44 68.41 61.54 54.76 115.26
SCI.c READ.c WRITE.c MATH.c SOC.cSCI.c 1.00 0.63 0.57 0.63 0.47READ.c 0.63 1.00 0.60 0.66 0.62WRITE.c 0.57 0.60 1.00 0.62 0.60MATH.c 0.63 0.66 0.62 1.00 0.54SOC.c 0.47 0.62 0.60 0.54 1.00
Example: High school and beyond dataWe're finally ready to fit our first CFA model using AMOS
Start by opening in the High School and Beyond dataset in SPSS
Click on the “Analyze” menu...down at the bottom should be an option for “Amos” -> Click it
Amos provides a very nice graphic interface for building our CFA models
Unidimensional CFA
Example: High school and beyond dataLet's build a single factor CFA model for the achievement tests
Notice that AMOS automatically adds the scale constraints
First path is constrained to 1.0Click on the “view” menu option and view the variables in the dataset
Drag the academic achievement tests to the desired box in the model
Click on the “plugins” menu optionUse the “name unobserved variables” to automatically name your latent variables for you
Before you estimate the parameters think about what you expect to see!!! (df, what estimates)
Unidimensional CFA
Example: High school and beyond dataClick on the “Analyze” menu option and then “Calculate the estimates”Examine the model with the estimates placed on itCheck your output window to make sure the model converged...”minimum was achieved”
Let's examine the local and global fit information
Does the model fit well?
Unidimensional CFA
Example: High school and beyond dataLet's fit the model with the latent variance constrained to 1.0
Select the latent variable and add the 1.0 constraintSelect the path and remove the 1.0 constraint.
Does the model fit change?Do the parameter estimates change?
Unidimensional CFA
Local FitAre the regression slopes all significantly difference from zero in the populationLook at the t-tests
Global FitChi-square test
Can be overpowered (too sensitive) with sample sizes over 200Sensitive to nonnormality
Goodness of fit indices
Global FitComparative fit index
Indicates reduction in model misfit of a target model relative to a baseline (independence) model
.90-.95 = acceptable; above .95 is good
Goodness of fit indices
CFI =null
2 df null model2 df model
model2 df model
Global FitRoot Mean Square Error of Approximation
RMSEA
Discrepancy per degree of freedomRMSEA ≤ 0.05 → Close fit0.05 <RMSEA ≤ 0.08 → Reasonable fitRMSEA > 0.1 → Poor fit
Goodness of fit indices
RMSEA= model2 /df model1
N 1
Global FitStandardized Root Mean Square Residual
SRMRStandardized difference between the observed covariance matrix and predicted covariance matrixA value of zero indicates perfect fitThis measure tends to be smaller as sample size increases and as the number of parameters in the model increases
SRMR < .08 is good fit
Goodness of fit indices
Global FitHu & Bentler
SRMR ≤ .08 AND (CFI ≥ .95 OR RMSEA ≤ .06)
Goodness of fit indices
ReliabilityThe reliability of a set of indicators as a measure of the latent variable may be determined from the estimated parameters.If the indicators are homogeneous (i.e. ,unidimensional)
Reliability is the ratio of true score variance to total variance
Unidimensional CFA
( )( )
2
2 2i
i
i δ
λω
λ σ=
+
∑∑ ∑
Omega is:Ratio of variance due to the common attribute to the total variance in X
Square of the correlation between X (observed score) and the the common factor
Correlation between two parallel test scores
Square of the correlation between the total score of m indicators and the total score of an infinite set of indicators
Unidimensional CFA
Parallel IndicatorsEqual lambdas (loadings)Equal error variances
Tau-Equivalent IndicatorsEqual lambdas
Congeneric IndicatorsNothing needs to be equal
Can compare model fit to determine the best representation
Unidimensional CFA
If the tau-equivalent model holds!Omega = Coefficient alpha
If not tau-equivalent, alpha will be less than omega
Standards for reliability (Nunnaly,1968).7 for research.9 for application
.7 has become the standard
Unidimensional CFA
Models are nested whenever one model is the same as the other model except that it has some parameters constrainedIn other words, if you unconstrained some parameters in one model, you would get the otherWhen models are nested, we can perform a Chi-square difference test to compare them
Nested Model Comparisons
Examples:Testing for redundancy - A one factor model is nested within a two factor simple structure CFA (constrains correlation between factors to equal 1)Testing for orthogonality - an orthogonal factor structure is nested within an oblique structureTesting for tau equivalence - a model with factor loadings constrained to be equal is nested within one that lets all loadings be different.
Nested Model Comparisons
Chi-Square difference testH0: More restrictive model (the one with some parameters constrained)HA: Less restrictive modelChi-square test to compare these models = Chi-square test value for More restrictive model - Chi-square test value for Less restrictive modelCompare this test stat to a chi-squared distribution with d.f. = (d.f. for More restrictive model) - (d.f. for Less restrictive model
Nested Model Comparisons
Example: Tau-equivalent modelH0: Tau-equivalent model vs HA: Congeneric modelH_0: 1 = a1_1 = a2_1 = a3_1 = a4_1 = a5_vs HA: all factor loadings can be differentChi-square test value for Congeneric model = 238.5 with 9 degrees of freedomChi-square test value for Tau-equivalent model = 263.8 with 14 degrees of freedomChi-square difference test for hypothesis above = 263.8-238.5 = 25.3 with 14-9 = 5 dfThe p-value associated with this test is < .0001, so we reject the Null hypothesis in favor of the alternative.Thus the Tau-equivalent model (more restrictive
Nested Model Comparisons
Chi-squared test of model fit and standard errors for factor loadings are based on normality assumption for x|f and f and asymptotic theoryHow big should n be before we trust these numbers????
∗n > 15 (number of free parameters)How non-normal is non-normal??Rules of thumb - examine univariate skew and kurtosis. West, Finch and Curran (1995) “Structural equation models with non-normal variables: Problems and remedies”. In R.H. Hoyle (Ed.) Structural equation modeling (pp. 56-75) Thousand Oaks, CA: Sage.
Sample Size and Normality
d
Missing Data
d
Factor Scores