12.11.2012 Views

growth curves and extensions using mplus - UCLA Integrated ...

growth curves and extensions using mplus - UCLA Integrated ...

growth curves and extensions using mplus - UCLA Integrated ...

SHOW MORE
SHOW LESS

Transform your PDFs into Flipbooks and boost your revenue!

Leverage SEO-optimized Flipbooks, powerful backlinks, and multimedia content to professionally showcase your products and significantly increase your reach.

GROWTH CURVES AND EXTENSIONS USING MPLUS<br />

Alan C. Acock<br />

alan.acock@oregonstate.edu<br />

Department of HDFS<br />

322 Milam Hall<br />

Oregon State University<br />

Corvallis, OR 97331<br />

7/2008<br />

This document <strong>and</strong> selected references, data, <strong>and</strong> programs can be downloaded from<br />

http://oregonstate.edu/~acock/<strong>growth</strong><br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 1


GROWTH CURVES AND EXTENSIONS USING MPLUS<br />

Outline<br />

Topic Page<br />

1 Brief summary of topics 3<br />

2 A <strong>growth</strong> curve 4<br />

3 Quadratic terms in <strong>growth</strong> <strong>curves</strong> 17<br />

4 An alternative—developmental time 23<br />

5 Working with missing values 24<br />

6 Multiple cohort <strong>growth</strong> model with missing waves 25<br />

7 Multiple group models with <strong>growth</strong> <strong>curves</strong> 27<br />

8 Alternative to multiple group analysis 35<br />

9 Growth <strong>curves</strong> with time invariant covariates 41<br />

10 Mediational models with time invariant covariates 49<br />

11 Time varying covariates 50<br />

12 References 51<br />

Goal of the Workshop<br />

The goal of this workshop is to explore a variety of applications of latent <strong>growth</strong> curve models<br />

<strong>using</strong> the Mplus program. Because we will cover a wide variety of applications <strong>and</strong> <strong>extensions</strong> of<br />

<strong>growth</strong> curve modeling, we will not cover each of them in great detail. At the end of this<br />

workshop it is hoped that participants will be able to run Mplus programs to execute a variety of<br />

<strong>growth</strong> curve modeling applications <strong>and</strong> to interpret the results correctly.<br />

Assumed Background<br />

Participants should be familiar with the content in Introduction to Mplus that is located at<br />

www.oregonstate.edu/~acock/<strong>growth</strong> . It will be assumed that participants in the workshop have<br />

some background in Structural Equation Modeling. Background in multilevel analysis will also be<br />

useful, but is not assumed. It is possible to learn how to estimate the specific models we will cover<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 2


without a comprehensive knowledge of Mplus, but some background <strong>using</strong> an SEM program is<br />

useful.<br />

Introduction to Growth Curve Modeling<br />

1 Brief Summary of Topics<br />

Growth Curves are a new way of thinking that is ideal for longitudinal studies. Instead of<br />

predicting a person’s score on a variable (e.g., mean comparison among scores at different time<br />

points or relationships among variables at different time points), we predict their <strong>growth</strong><br />

trajectory—what is their level on the variable AND how is this changing. We will present a<br />

conceptual model, show how to apply the Mplus program, <strong>and</strong> interpret the results. Once we can<br />

estimate <strong>growth</strong> trajectories, the more interesting issues of explaining individual differences in<br />

trajectories (why some people go up, down, or stay the same). More advanced topics we will<br />

introduce include:<br />

1. Growth Curves with Limited Outcome Variables<br />

Sometimes a researcher is interested in <strong>growth</strong> on variable that is not continuous.<br />

• It may be a binary variable.<br />

o Ever drinking alcohol for adolescents.<br />

o Retention of program participants across 10 sessions.<br />

• Some times a researcher is interested in a count variable that involves a relatively rare<br />

event<br />

o Number of days an adolescent has 5+ drinks of alcohol in the last 30 days.<br />

o Number of times a couple has severe verbal conflict in last week.<br />

• Sometimes we are interested in both types of variables at the same time.<br />

o Onset of sexual intercourse between 12 – 18 cf. frequency of sexual intercourse.<br />

o Onset of binge drinking cf. frequency of binge drinking.<br />

o Different variables may predict the binary aspect (onset) of the variable than<br />

predict the count aspect of the variable.<br />

• These applications can be extremely CPU intensive, taking hours to converge. Mplus<br />

utilizes multiple processors.<br />

2. Growth Mixture Models<br />

It is possible to use Mplus to do an exploratory <strong>growth</strong> curve analysis where our focus is on<br />

the person <strong>and</strong> not the variable. We can locate clusters of people who share similar <strong>growth</strong><br />

trajectories. This is exploratory research <strong>and</strong> the st<strong>and</strong>ards for it are still evolving.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 3


• A study of alcohol consumption from age 15 to 30. It is possible to empirically identify<br />

different clusters of people.<br />

o One cluster may never drink or never drink very much.<br />

o A second cluster may have increasing alcohol consumption up to about 22 or 23<br />

<strong>and</strong> then a gradual decline.<br />

o A third cluster may be very similar to the second cluster but not decline after 23.<br />

• A study of marital conflict across the first 5 years of marriage<br />

o One cluster may have persistently low conflict<br />

o One cluster may have monotonic escalating conflict<br />

o One cluster may have persistently high conflict<br />

o One cluster may have initially low conflict, peak at the end of the first year, <strong>and</strong><br />

then have monotonically declining conflict.<br />

After deriving these clusters of people who share <strong>growth</strong> trajectories, it is possible to compare<br />

them to find what differentiates membership in the different clusters.<br />

• What covariates predict cluster membership?<br />

• What outcome variables does cluster membership predict?<br />

2 A Simple Growth Curve<br />

Estimating a basic <strong>growth</strong> curve <strong>using</strong> Mplus is quite easy. When developing a complex model it<br />

is best to start easy <strong>and</strong> gradually build complexity.<br />

• Starting easy should include data screening to evaluate the distributions of the variables,<br />

patterns of missing values, <strong>and</strong> possible outliers.<br />

• Even if you have a theoretically specified model that is complex, always start with the<br />

simplest model <strong>and</strong> gradually add the complexity.<br />

• Here we will show how structural equation modeling conceptualizes a latent <strong>growth</strong> <strong>curves</strong>.<br />

Before showing a figure to represent a <strong>growth</strong> curve, we examine a small sample of our<br />

observations:<br />

• Data is from the National Longitudinal Survey of Youth that started in 1997.<br />

• We use the cohort that was 12 years old in 1997 <strong>and</strong> examine their trajectory for the BMI.<br />

• Some may not like <strong>using</strong> the BMI on this age group, but this is only to illustrate an<br />

application of <strong>growth</strong> curve modeling.<br />

• The following graph of 10 r<strong>and</strong>omly selected kids was produced by Mplus<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 4


• A BMI value of 25 is considered overweight <strong>and</strong> a BMI of 30 is considered obese (I’m<br />

aware of problems with the BMI as a measure of obesity <strong>and</strong> with its limitations when used<br />

for adolescents)<br />

• With just 10 observations it is hard to see much of a trend, but it looks like people are<br />

getting a bigger BMI score as they get older.<br />

• The X-axis value of 0 is when the adolescent was 12 years old; the 1 is when the adolescent<br />

was 13 years old, etc. We are <strong>using</strong> seven waves of data (labeled 0 to 6) from the panel<br />

study.<br />

A <strong>growth</strong> curve requires us to have a model <strong>and</strong> we should draw this before writing the Mplus<br />

program. Figure 1 shows a model for our simple <strong>growth</strong> curve:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 5


This figure is much simpler than it first appears.<br />

• The key variables are the two latent variables labeled the Intercept <strong>and</strong> the Slope.<br />

• The Intercept<br />

a. The intercept represents the initial level <strong>and</strong> is sometimes called the initial level for this<br />

reason. It is the estimated initial level <strong>and</strong> its value may differ from the actual mean for<br />

BMI97 because in this case we are imposing a linear <strong>growth</strong> model.<br />

b. It may differ from the mean of BMI97 when covariates are added, especially when a zero<br />

value on covariates is rare <strong>and</strong> covariates are not centered (household income)<br />

c. Unless the covariates are centered, it usually makes sense to just call it an intercept rather<br />

than the initial level.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 6


d. The intercept is identified by the constant loadings of 1.0 going to each BMI score. Some<br />

programs call the intercept the constant, representing the constant effect to which other<br />

effects are added.<br />

e. It is possible to shift the intercept by how the waves are coded, e.g., -2 -1 0 1 2.<br />

• The slope<br />

a. Is identified by fixing the values of the paths to each BMI variable. In a publication you<br />

normally would not show the path to BMI97, since this is fixed at 0.0.<br />

b. We fix the other paths at 1.0, 2,0, 3.0, 4.0, 5.0, <strong>and</strong> 6.0. Where did we get these values? The<br />

first year is the base year or year zero. The BMI was measured each subsequent year so<br />

these are scored 1.0 through 6.0.<br />

c. Other values are possible. Suppose the survey was not done in 2000 or 2001 so that we had<br />

5 time points rather than 7. We would use paths of 0.0, 1.0, 2.0, 5.0, <strong>and</strong> 6.0 for years 1997,<br />

1998, 1997, 2002, <strong>and</strong> 2003, respectively.<br />

d. It is also possible to fix the first couple years <strong>and</strong> then allow the subsequent waves to be<br />

free.<br />

- This might make sense for a developmental process where the yearly intervals may not<br />

reflect the developmental rate. Developmental time may be quite different than<br />

chronological time.<br />

- This has the effect of “stretching” or “shrinking” time to the pattern of the data (Curran<br />

& Hussong, 2003).<br />

- An advantage of this approach is that it uses fewer degrees of freedom than adding a<br />

quadratic slope.<br />

- Mplus has a feature that allows each participant to have a different interval which is<br />

important when the time between waves varies.<br />

• Residual Variance <strong>and</strong> R<strong>and</strong>om Effects<br />

a. The individual variation around the Intercept <strong>and</strong> Slope are represented in Figure 1 by the RI<br />

<strong>and</strong> R2. These are the variance in the intercept <strong>and</strong> slope around their respective means.<br />

b. We expect substantial variance in both of these as some individuals have a higher or lower<br />

starting BMI <strong>and</strong> some individuals will increase (or decrease) their BMI at a different rate<br />

than the average <strong>growth</strong> rate.<br />

c. In addition to the mean intercept <strong>and</strong> slope, each individual will have their own intercept<br />

<strong>and</strong> slope. We say the intercept <strong>and</strong> the slope are r<strong>and</strong>om effects since they may vary<br />

across individuals.<br />

- They are r<strong>and</strong>om in the sense that each individual may have a steeper or flatter slope<br />

than the mean slope <strong>and</strong><br />

- Each individual may have a higher or lower initial level than the mean intercept.<br />

d. In our sample of 10 individuals shown above, notice one adolescent starts with a BMI<br />

around 12 <strong>and</strong> three adolescents start with a BMI around 30. Some children have a BMI that<br />

increases <strong>and</strong> others do not.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 7


e. The variances, RI <strong>and</strong> R2 are critical if we are going to explore more complex models with<br />

covariates (e.g., gender, psychological problems, race, household income, physical activity)<br />

that might explain why some individuals have a steeper or less steep <strong>growth</strong> rate than the<br />

average.<br />

• The ei terms represent individual error terms for each year. Some years may move above or<br />

below the <strong>growth</strong> trajectory described by our Intercept <strong>and</strong> Slope.<br />

• Sometimes it might be important to allow error terms to be correlated, especially subsequent<br />

pairs such as e97-e98, e98-e99, etc.<br />

Here is the Mplus program for a simple <strong>growth</strong> model:<br />

Title: bmi_<strong>growth</strong>.inp<br />

Basic <strong>growth</strong> curve<br />

Data:<br />

File is "C:\Mplus examples\bmi_stata.dat" ;<br />

Analysis: Processors = 2;<br />

Variable:<br />

Names are<br />

id grlprb_y boyprb_y grlprb_p boyprb_p<br />

male race_eth bmi97 bmi98 bmi99 bmi00<br />

bmi01 bmi02 bmi03 white black hispanic<br />

asian other;<br />

Missing are all (-9999) ;<br />

! Notice usevariables is limited to bmi variables<br />

Usevariables are bmi97 bmi98 bmi99 bmi00<br />

bmi01 bmi02 bmi03 ;<br />

Model:<br />

i s | bmi97@0 bmi98@1 bmi99@2 bmi00@3<br />

bmi01@4 bmi02@5 bmi03@6;<br />

Output:<br />

Sampstat Mod(3.84);<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00 bmi01<br />

bmi02 bmi03(*);<br />

What is new compared to an SEM program?<br />

• Usevariables are: subcomm<strong>and</strong> to only include the BMI variables since we are doing a<br />

<strong>growth</strong> curve for these variables.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 8


• We drop the Analysis: section if we have a single processor because we are doing basic<br />

<strong>growth</strong> curve <strong>and</strong> can use the default options. With multiple processors, this is included to tell<br />

Mplus how many processors to utilize. (This model ran 40% faster on a 2 processor IMac<br />

running Windows under Parallels.)<br />

• We have a Model: section because we need to describe the model. Mplus was designed after<br />

<strong>growth</strong> <strong>curves</strong> were well understood. There is a single line to describe our model:<br />

i s | bmi97@0 bmi98@1 bmi99@2 bmi00@3 bmi01@4 bmi02@5 bmi03@6;<br />

a. In this line the “i” <strong>and</strong> “s” st<strong>and</strong> for the intercept <strong>and</strong> slope, respectively. We could have<br />

called these anything such as intercept <strong>and</strong> slope or initial <strong>and</strong> trend. The<br />

vertical line, | (sometimes called “oar bar”, tells Mplus that it is about to define an<br />

intercept <strong>and</strong> slope.<br />

b. Defaults<br />

- The intercept is defined by a constant of 1.0 for each bmi variable. Intercept�bmij path<br />

is 1.0. Therefore, we do not need to mention this.<br />

- The slope is defined by fixing the path from the slope to bmi97 at 0, the path to bmi98<br />

at 1, etc. The @ sign is used for “at.” Don’t forget the semi-colon to end the comm<strong>and</strong>.<br />

- Mplus assumes that there is a residual variance for both the intercept <strong>and</strong> slope (RI <strong>and</strong><br />

R2) <strong>and</strong> that these covary. Therefore, we do not need to mention this<br />

- Mplus assumes there is uncorrelated r<strong>and</strong>om error, ei for each observed variable<br />

c. To allow e97 <strong>and</strong> e98 to be correlated, we would need to add a line saying bmi97 with<br />

bmi98; .<br />

- This may seem strange because we are not really correlating bmi97 with bmi98, but<br />

e97 with e98. Mplus knows this <strong>and</strong> we do not need to generate a separate set of names for<br />

the error terms.<br />

The last additional section in our Mplus program is for selecting what output we want Mplus to<br />

provide. There are many optional outputs of the program <strong>and</strong> we will only illustrate a few of these.<br />

The Output: section has the following lines<br />

Output:<br />

Sampstat Mod(3.84);<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00<br />

bmi01 bmi02 bmi03(*);<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 9


• The first line, Sampstat Mod(3.84) asks for sample statistics <strong>and</strong> modification indices for<br />

parameters we might free, as long as doing so would reduce chi-square by 3.84 (corresponding<br />

to the .05 level). We do not bother with parameter estimates that would have less effect than<br />

this. The default value is 10.0.<br />

• Next comes the Plot: subcomm<strong>and</strong> <strong>and</strong> we say that we want Type is Plot3; for our<br />

output. This gives us the descriptive statistics <strong>and</strong> graphs for the <strong>growth</strong> curve.<br />

• The last line of the program specifies the series to plot. By entering the variables with an (*)<br />

at the end we are setting a path at 0.0 for bmi97, 1.0 for bmi98, etc.<br />

Annotated Selected Growth Curve Output<br />

The following is selected output with comments:<br />

Mplus VERSION 5.1<br />

MUTHEN & MUTHEN<br />

07/01/2008 2:40 PM<br />

*** WARNING<br />

Data set contains cases with missing on all variables.<br />

These cases were not included in the analysis.<br />

Number of cases with missing on all variables: 3<br />

1 WARNING(S) FOUND IN THE INPUT INSTRUCTIONS<br />

Mplus uses all available data assuming MAR. There were three cases that were<br />

dropped because they had no BMI report for any wave.<br />

bmi_<strong>growth</strong>.inp<br />

Basic <strong>growth</strong> curve<br />

SUMMARY OF ANALYSIS<br />

Number of groups 1<br />

Number of observations 1768<br />

With listwise deletion we would have an N = 1102<br />

Number of dependent variables 7<br />

Number of independent variables 0<br />

Number of continuous latent variables 2<br />

Observed dependent variables<br />

Continuous<br />

BMI97 BMI98 BMI99 BMI00 BMI01 BMI02<br />

BMI03<br />

Continuous latent variables<br />

I S<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 10


The following is a very nice analysis of patterns of missing values.<br />

Estimator ML<br />

Information matrix OBSERVED<br />

SUMMARY OF DATA<br />

Number of missing data patterns 81<br />

COVARIANCE COVERAGE OF DATA<br />

Minimum covariance coverage value 0.100<br />

PROPORTION OF DATA PRESENT<br />

Covariance Coverage<br />

BMI97 BMI98 BMI99 BMI00 BMI01<br />

________ ________ ________ ________ ________<br />

BMI97 0.925<br />

BMI98 0.847 0.902<br />

BMI99 0.850 0.856 0.910<br />

BMI00 0.842 0.846 0.864 0.906<br />

BMI01 0.839 0.837 0.854 0.859 0.904<br />

BMI02 0.796 0.794 0.805 0.811 0.817<br />

BMI03 0.777 0.775 0.788 0.788 0.801<br />

Covariance Coverage<br />

BMI02 BMI03<br />

________ ________<br />

BMI02 0.861<br />

BMI03 0.774 0.840<br />

Check means to see if there is a clear overall trajectory<br />

SAMPLE STATISTICS<br />

ESTIMATED SAMPLE STATISTICS<br />

Means<br />

BMI97 BMI98 BMI99 BMI00 BMI01<br />

________ ________ ________ ________ ________<br />

1 20.572 21.839 22.651 23.305 23.846<br />

Means<br />

BMI02 BMI03<br />

________ ________<br />

1 24.390 24.935<br />

Correlations<br />

BMI97 BMI98 BMI99 BMI00 BMI01<br />

________ ________ ________ ________ ________<br />

BMI97 1.000<br />

BMI98 0.764 1.000<br />

BMI99 0.765 0.850 1.000<br />

BMI00 0.721 0.812 0.853 1.000<br />

BMI01 0.709 0.799 0.853 0.856 1.000<br />

BMI02 0.652 0.720 0.745 0.752 0.813<br />

BMI03 0.651 0.707 0.737 0.751 0.815<br />

Correlations<br />

BMI02 BMI03<br />

________ ________<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 11


BMI02 1.000<br />

BMI03 0.766 1.000<br />

TESTS OF MODEL FIT<br />

Chi-Square Test of Model Fit<br />

Value 268.041<br />

Degrees of Freedom 23<br />

P-Value 0.0000<br />

Chi-Square Test of Model Fit for the Baseline Model<br />

Value 11502.912<br />

Degrees of Freedom 21<br />

P-Value 0.0000<br />

CFI/TLI<br />

CFI 0.979<br />

TLI 0.981<br />

Loglikelihood<br />

H0 Value -27739.720<br />

H1 Value -27605.699<br />

Information Criteria<br />

Number of Free Parameters 12<br />

Akaike (AIC) 55503.439<br />

Bayesian (BIC) 55569.171<br />

Sample-Size Adjusted BIC 55531.048<br />

(n* = (n + 2) / 24)<br />

RMSEA (Root Mean Square Error Of Approximation)<br />

Estimate 0.078<br />

90 Percent C.I. 0.069 0.086<br />

Probability RMSEA


BMI98 1.000 0.000 999.000 999.000<br />

BMI99 2.000 0.000 999.000 999.000<br />

BMI00 3.000 0.000 999.000 999.000<br />

BMI01 4.000 0.000 999.000 999.000<br />

BMI02 5.000 0.000 999.000 999.000<br />

BMI03 6.000 0.000 999.000 999.000<br />

Intercept <strong>and</strong> slope have significant covariance<br />

S WITH<br />

I 0.408 0.073 5.559 0.000<br />

Means<br />

I 21.035 0.100 210.352 0.000<br />

S 0.701 0.017 40.663 0.000<br />

Growth curve is BMI’ = 21.035 + .701×Year<br />

Intercepts<br />

BMI97 0.000 0.000 999.000 999.000<br />

BMI98 0.000 0.000 999.000 999.000<br />

BMI99 0.000 0.000 999.000 999.000<br />

BMI00 0.000 0.000 999.000 999.000<br />

BMI01 0.000 0.000 999.000 999.000<br />

BMI02 0.000 0.000 999.000 999.000<br />

BMI03 0.000 0.000 999.000 999.000<br />

Variances<br />

I 15.051 0.597 25.209 0.000<br />

S 0.255 0.018 14.228 0.000<br />

There is a big r<strong>and</strong>om intercept <strong>and</strong> r<strong>and</strong>om slope effect. The st<strong>and</strong>ard<br />

deviation is sqrt(.255) = .50. Putting plus or minus two st<strong>and</strong>ard deviations<br />

around the slope of .70 shows how big the variance is. The st<strong>and</strong>ard<br />

deviation for the intercept is sqrt(15.051) = 3.880. BMI is probably skewed<br />

positively.<br />

Residual Variances<br />

BMI97 5.730 0.268 21.413 0.000<br />

BMI98 3.276 0.164 19.942 0.000<br />

BMI99 3.223 0.146 22.009 0.000<br />

BMI00 4.361 0.185 23.538 0.000<br />

BMI01 2.845 0.150 19.005 0.000<br />

BMI02 9.380 0.397 23.622 0.000<br />

BMI03 8.589 0.422 20.345 0.000<br />

QUALITY OF NUMERICAL RESULTS<br />

Condition Number for the Information Matrix 0.656E-02<br />

(ratio of smallest to largest eigenvalue)<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 13


MODEL MODIFICATION INDICES<br />

Minimum M.I. value for printing the modification index 3.840<br />

We don’t want to change the intercept loadings of 1.0. We might think about<br />

a nonlinear <strong>growth</strong>. We might think about correlating adjacent error terms.<br />

The suggested correlation between E1 <strong>and</strong> E7 indicates a straight line is<br />

missing both ends, hence a curve of some kind? Muthen suggests to try not to<br />

mess with the intercepts.<br />

M.I. E.P.C. Std E.P.C. StdYX E.P.C.<br />

BY Statements<br />

I BY BMI97 112.472 -0.038 -0.147 -0.032<br />

I BY BMI98 6.440 0.007 0.027 0.006<br />

I BY BMI99 33.234 0.014 0.054 0.012<br />

I BY BMI00 13.026 0.010 0.037 0.008<br />

I BY BMI02 4.015 -0.008 -0.032 -0.005<br />

I BY BMI03 28.212 -0.023 -0.091 -0.015<br />

S BY BMI97 70.828 -0.825 -0.417 -0.091<br />

S BY BMI99 18.208 0.276 0.139 0.030<br />

S BY BMI00 8.062 0.204 0.103 0.021<br />

S BY BMI03 38.314 -0.755 -0.382 -0.062<br />

WITH Statements<br />

BMI99 WITH BMI98 12.747 0.449 0.449 0.138<br />

BMI00 WITH BMI97 9.699 -0.511 -0.511 -0.102<br />

BMI00 WITH BMI99 26.084 0.641 0.641 0.171<br />

BMI01 WITH BMI97 6.914 -0.388 -0.388 -0.096<br />

BMI01 WITH BMI98 11.566 -0.403 -0.403 -0.132<br />

BMI01 WITH BMI00 5.456 0.310 0.310 0.088<br />

BMI02 WITH BMI97 8.645 0.715 0.715 0.098<br />

BMI02 WITH BMI99 9.066 -0.544 -0.544 -0.099<br />

BMI02 WITH BMI00 9.560 -0.633 -0.633 -0.099<br />

BMI03 WITH BMI97 37.342 1.564 1.564 0.223<br />

BMI03 WITH BMI99 22.526 -0.874 -0.874 -0.166<br />

BMI03 WITH BMI00 11.717 -0.724 -0.724 -0.118<br />

BMI03 WITH BMI02 11.053 1.083 1.083 0.121<br />

Means/Intercepts/Thresholds<br />

[ BMI97 ] 97.476 -0.754 -0.754 -0.165<br />

[ BMI98 ] 7.230 0.155 0.155 0.035<br />

[ BMI99 ] 25.098 0.257 0.257 0.056<br />

[ BMI00 ] 10.542 0.185 0.185 0.038<br />

[ BMI02 ] 4.646 -0.189 -0.189 -0.032<br />

[ BMI03 ] 22.536 -0.448 -0.448 -0.073<br />

There are a number of plots available. These are not bad, but Stata or some<br />

other package, even Excel, could do nicer graphs.<br />

PLOT INFORMATION<br />

The following plots are available:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 14


Histograms (sample values, estimated factor scores, estimated values)<br />

Scatterplots (sample values, estimated factor scores, estimated values)<br />

Sample means<br />

Estimated means<br />

Sample <strong>and</strong> estimated means<br />

Observed individual values<br />

Estimated individual values<br />

Here are Some of the Available Plots<br />

It is often useful to show the actual means for a small r<strong>and</strong>om sample of participants. These are<br />

Sample Means.<br />

• Click on Graphs<br />

• Observed Individual Values<br />

This gives you a menu where you can make some selections. I used the clock to seed a r<strong>and</strong>om<br />

generation of observations.<br />

Here I selected R<strong>and</strong>om Order <strong>and</strong> for 20 cases. This results in the following graph:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 15


This shows one person who started at an obese BMI = 30 <strong>and</strong> then dropped down. However, most<br />

people increased gradually.<br />

Next, let’s look at a plot of the actual means <strong>and</strong> the estimated means <strong>using</strong> our linear <strong>growth</strong><br />

model. Click on<br />

• Graphs <strong>and</strong> then select View graphs<br />

• Sample <strong>and</strong> estimated means.<br />

• Demonstrate how to edit the graph.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 16


Notice that there is a clear <strong>growth</strong> trend in BMI. A BMI of 15-20 is considered healthy <strong>and</strong> a BMI<br />

of 25 is considered overweight. Notice what happens to American youth between the age of 12<br />

<strong>and</strong> the age of 18.<br />

3 A Growth Curve with a Quadratic Term<br />

This graph is useful to seeing if there is a nonlinear trend. Changing the scale of the Y-axis can<br />

clarify this. It is simple to add a quadratic term, if the curve is departing from linearity.<br />

• Looking at the graph it may seem that the linear trend works very well, but our RMSEA<br />

was a bit big.<br />

• The estimated initial BMI is higher than the observed mean.<br />

• The estimated BMI at 2003 is also higher than the observed mean<br />

• A quadratic might pick this up by having a curve that drops slightly to pick up the BMI97<br />

mean <strong>and</strong> the BMI2003 mean.<br />

• Estimation of three terms (Intercept, Lin1ear trend, Quadratic trend) requires at least four<br />

waves of data, but more waves are highly desirable for a good test of the quadratic term.<br />

The conceptual model in Figure 1 will be unchanged except a third latent variable is added.<br />

• We will have the Intercept, Slope, now called linear trend, <strong>and</strong> the new latent variable<br />

called the Quadratic trend.<br />

• Like the first two, the Quadratic trend will have a residual variance (R3) that will freely<br />

correlated with R1 <strong>and</strong> R2.<br />

• The paths from the quadratic trend to the individual BMI variables will be the square of<br />

the path from the Linear trend to the BMI variables. Hence<br />

a. The values for the linear trend will remain 0.0, 1.0, 2.0, 3.0, 4.0, 5.0, <strong>and</strong> 6.0.<br />

b. For the quadratic these values will be 0.0, 1.0, 4.0, 9.0, 16.0, 25.0, <strong>and</strong> 36.0.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 17


You really appreciate the defaults in Mplus when you see what we need to change in the Mplus<br />

program when we add a quadratic slope. Here is the only change we need to make:<br />

Model:<br />

i s q| bmi97@0 bmi98@1 bmi99@2 bmi00@3 bmi01@4 bmi02@5 bmi03@6;<br />

Mplus will know that the quadratic, q (we could use any name) will have values that are the<br />

square of the values for the slope, s.<br />

Title: bmi_guadratic.inp<br />

Quadratic <strong>growth</strong> curve<br />

Data:<br />

File is "C:\Mplus examples\bmi_stata.dat" ;<br />

Variable:<br />

Names are<br />

id grlprb_y boyprb_y grlprb_p boyprb_p<br />

male race_eth bmi97 bmi98 bmi99 bmi00<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 18


mi01 bmi02 bmi03 white black hispanic<br />

asian other;<br />

Missing are all (-9999) ;<br />

! usevariables is limited to bmi variables<br />

Usevariables are bmi97 bmi98 bmi99 bmi00<br />

bmi01 bmi02 bmi03 ;<br />

Model:<br />

i s q| bmi97@0 bmi98@1 bmi99@2 bmi00@3<br />

bmi01@4 bmi02@5 bmi03@6;<br />

Output:<br />

Sampstat Mod(3.84);<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00 bmi01<br />

bmi02 bmi03(*);<br />

Here is selected output:<br />

TESTS OF MODEL FIT<br />

We had 23 degrees of freedom with the linear <strong>growth</strong> curve <strong>and</strong> a chi-square<br />

of 268.041. Now we have 19 degrees of freedom <strong>and</strong> a chi-square of 73.121.<br />

Where did we lose four degrees of freedom?<br />

• Mean for the quadratic<br />

• Variance of the quadratic<br />

• Covariance of quadratic residual with intercept residual<br />

• Covariance of quadratic residual with slope residual<br />

Did we improve our fit?<br />

• 268.041-73.121 = 194.92 with 4 degrees of freedom, p <<br />

.001<br />

Does our model fit?<br />

• Chi-square (19) = 73.121, p < .001, but<br />

• CFI = .995<br />

• RMSEA = .040<br />

Chi-Square Test of Model Fit<br />

Value 73.121<br />

Degrees of Freedom 19<br />

P-Value 0.0000<br />

Chi-Square Test of Model Fit for the Baseline Model<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 19


CFI/TLI<br />

Loglikelihood<br />

Value 11502.912<br />

Degrees of Freedom 21<br />

P-Value 0.0000<br />

CFI 0.995<br />

TLI 0.995<br />

H0 Value -27642.260<br />

H1 Value -27605.699<br />

Information Criteria<br />

Number of Free Parameters 16<br />

Akaike (AIC) 55316.520<br />

Bayesian (BIC) 55404.161<br />

Sample-Size Adjusted BIC 55353.330<br />

(n* = (n + 2) / 24)<br />

RMSEA (Root Mean Square Error Of Approximation)<br />

Estimate 0.040<br />

90 Percent C.I. 0.031 0.050<br />

Probability RMSEA


Means<br />

I 20.713 0.101 204.728 0.000<br />

S 1.060 0.044 23.834 0.000<br />

Q -0.063 0.007 -8.585 0.000<br />

Variances<br />

I 14.273 0.638 22.382 0.000<br />

S 1.141 0.139 8.184 0.000<br />

Q 0.029 0.004 7.730 0.000<br />

Residual Variances<br />

BMI97 4.635 0.306 15.132 0.000<br />

BMI98 3.340 0.162 20.643 0.000<br />

BMI99 2.852 0.143 19.954 0.000<br />

BMI00 3.994 0.182 21.926 0.000<br />

BMI01 2.880 0.154 18.762 0.000<br />

BMI02 9.343 0.394 23.690 0.000<br />

BMI03 5.677 0.507 11.192 0.000<br />

MODEL MODIFICATION INDICES<br />

Minimum M.I. value for printing the modification index 3.840<br />

M.I. E.P.C. Std E.P.C. StdYX E.P.C.<br />

BY Statements<br />

I BY BMI97 24.292 -0.024 -0.090 -0.021<br />

I BY BMI98 9.860 0.008 0.032 0.007<br />

I BY BMI99 5.419 0.006 0.022 0.005<br />

I BY BMI01 12.777 -0.009 -0.035 -0.007<br />

I BY BMI03 14.857 0.024 0.092 0.016<br />

S BY BMI97 18.253 -0.363 -0.388 -0.089<br />

S BY BMI98 7.381 0.126 0.135 0.031<br />

S BY BMI01 9.168 -0.137 -0.147 -0.029<br />

S BY BMI03 10.308 0.349 0.373 0.063<br />

Q BY BMI97 12.444 3.442 0.589 0.136<br />

Q BY BMI99 4.868 -1.114 -0.191 -0.041<br />

Q BY BMI01 11.767 1.725 0.295 0.058<br />

Q BY BMI03 13.934 -5.610 -0.961 -0.163<br />

ON/BY Statements<br />

Q ON I /<br />

I BY Q 999.000 0.000 0.000 0.000<br />

WITH Statements<br />

BMI98 WITH BMI97 11.493 -1.044 -1.044 -0.265<br />

BMI99 WITH BMI98 8.019 0.361 0.361 0.117<br />

BMI01 WITH BMI98 8.978 -0.354 -0.354 -0.114<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 21


BMI02 WITH BMI01 12.482 0.694 0.694 0.134<br />

BMI03 WITH BMI02 5.261 -1.040 -1.040 -0.143<br />

Means/Intercepts/Thresholds<br />

[ BMI97 ] 23.635 -0.492 -0.492 -0.113<br />

[ BMI98 ] 11.403 0.191 0.191 0.043<br />

[ BMI01 ] 9.093 -0.166 -0.166 -0.032<br />

[ BMI03 ] 13.777 0.495 0.495 0.084<br />

PLOT INFORMATION<br />

The following plots are available:<br />

Histograms (sample values, estimated factor scores, estimated values)<br />

Scatterplots (sample values, estimated factor scores, estimated values)<br />

Sample means<br />

Estimated means<br />

Sample <strong>and</strong> estimated means<br />

Observed individual values<br />

Estimated individual values<br />

• The fit is so good because the estimated means <strong>and</strong> observed means are so close.<br />

• However, there is still significance variance (r<strong>and</strong>om effects for both the intercept <strong>and</strong> the<br />

slope) among individual adolescents that still needs to be explained.<br />

• Here are 20 estimated individual <strong>growth</strong> <strong>curves</strong>.<br />

a. Notice that each of these is a curve, but they start at different initial levels <strong>and</strong> have<br />

different trajectories.<br />

b. Next, we want to use covariates to explain these differences in the initial levels <strong>and</strong> <strong>growth</strong><br />

trajectories.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 22


4 An Alternative to Use of a Quadratic Slope<br />

An alternative to adding a quadratic slope is to allow some of the time loadings to be<br />

free.<br />

• We have used loadings of 0, 1, 2, 3, 4, 5, <strong>and</strong> 6 for the linear slope <strong>and</strong> 0, 1, 4,<br />

9, 16, 25, <strong>and</strong> 36 for the quadratic slope. Alternatively<br />

• We could allow all but two of the loadings to be free. We might use loadings of<br />

0, 1, *, *, *, * .<br />

• It is necessary to have the 0 <strong>and</strong> 1 fixed but the 1 does not have to be second;<br />

we could use 0, *, *, *, *,1.<br />

You may ask how you could justify allowing some of the time loadings to be free if<br />

there was a one month or one year difference between waves of data. The answer is<br />

that developmental time may be different than chronological time.<br />

Allowing these loadings to be free has an advantage over the quadratic in that it uses<br />

fewer degrees of freedom but still allows for <strong>growth</strong> spurts.<br />

This model is not nested under a quadratic, but you could think of a linear <strong>growth</strong><br />

model with fixed values for each year (0, 1, 2, 3, 4, 5, 6) being nested within the free<br />

model that uses 0, 1, *, *, *, *. If the free model fits much better than the fixed linear<br />

model, you might use this instead of the quadratic model.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 23


5 Working with Missing Values<br />

Mplus has two ways of working with missing values. The simplest is to use full information<br />

maximum likelihood estimation with missing values (FIML). This uses all available data <strong>and</strong> is<br />

the default. For example, some adolescents were interviewed all six years but others may have<br />

skipped one, two, or even more years. We use all available information with this approach. The<br />

second approach is to utilize multiple imputations.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 24


• Multiple imputations should not be confused with single imputation available from SPSS if a<br />

person purchases their missing values module <strong>and</strong> which gives incorrect st<strong>and</strong>ard errors.<br />

• Multiple imputation involves<br />

a. Imputing multiple datasets (usually 5-10) <strong>using</strong> appropriate procedures.<br />

b. Estimating the model for each of these datasets, <strong>and</strong><br />

c. Then pooling the estimates <strong>and</strong> st<strong>and</strong>ard errors.<br />

When the st<strong>and</strong>ard errors are pooled this way, they incorporate the variability across the 5-10<br />

solutions <strong>and</strong> are thereby produced unbiased estimates of st<strong>and</strong>ard errors. Multiple imputations<br />

can be done with:<br />

• Norm, a freeware program that works for normally distributed, continuous variables <strong>and</strong> is<br />

often used even on dichotomized variables.<br />

• A Stata user has written a program called ice that is an implementation of the S-Plus/R<br />

program called MICE, that has advantages over Norm. It does the imputation by <strong>using</strong> different<br />

estimation models for outcome variables that are continuous, counts, or categorical. See<br />

Royston (2005). SAS has similar capabilities.<br />

• Mplus can read these multiple datasets, estimate the model for each dataset, <strong>and</strong> pool the<br />

estimates <strong>and</strong> their st<strong>and</strong>ard errors.<br />

We will not illustrate the multiple imputation approach because that involves working with other<br />

programs to impute the datasets. However, the Mplus User’s Guide, discusses how you specify the<br />

datasets in the Data: section.<br />

6 Multiple Cohort Growth Model with Missing Waves<br />

Major datasets often have multiple cohorts. NLSY97 has youth who were 12-18 in 1997. Seven<br />

years later, they are 19-25. It is quite likely that many <strong>growth</strong> processes that involve going from<br />

the age of 12 to the age of 19 are different than going from 19-25. For example, involvement in<br />

minor crimes (petty theft, etc.) may increase from 12 to 19, but then decrease from there to 25.<br />

Here is what we might have for our NLSY97 data (data inside tables are scores, person 1, born in<br />

1985, in 1997 at age of 12 had a score of 3 on the outcome variable)<br />

Score by survey year for a single case from each cohort<br />

Survey Year<br />

Individual Cohort 1997 1998 1999 2000 2001 2002 2003<br />

1 1985 3 4 5 6 7 7 8<br />

2 1985 2 4 3 5 6 7 7<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 25


3 1984 4 5 6 7 6 6 5<br />

4 1982 6 7 5 4 3 2 2<br />

5 1982 5 5 6 4 2 2 1<br />

We can rearrange this data<br />

Data for first 5 cases<br />

Case Cohort 12 13 14 15 16 17 18 19 20 21<br />

1 1985 3 4 5 6 7 7 8 * * *<br />

2 1985 2 4 3 5 6 7 7 * * *<br />

3 1984 * 4 5 6 7 6 6 5 * *<br />

4 1982 * * * 6 7 5 4 3 2 2<br />

5 1982 * * * 5 5 6 4 2 2 1<br />

• In this table the top row is the age at which the data was collected. To capture everybody we<br />

would need to extend the table to HD25 because the youth who were 18 in 1997 are 25 seven<br />

years latter.<br />

• This table would have massive amounts of missing data, but the missingness would not be<br />

related to other variables. It would be missing completely at r<strong>and</strong>om (MCAR).<br />

• We could develop a <strong>growth</strong> curve that covered the full range from age 12 to age 25. We would<br />

have 14 waves of data even though each participant was only measured 7 times. Each<br />

participant would have data for a maximum of 7 of the years <strong>and</strong> have missing values for a<br />

minimum of 7 years.<br />

• We would want to estimate a <strong>growth</strong> model with a quadratic term <strong>and</strong> expect the linear slope to<br />

be positive (<strong>growth</strong> from 12-18) <strong>and</strong> the quadratic term to be negative (decline from 18-25).<br />

• Mplus has a special Analysis: type called MCOHORT. There is an example on the Mplus<br />

WebPage <strong>and</strong> we will not cover it here. This is an extraordinary way to deal with missing<br />

values.<br />

Here is an example from data Muthén analyzed:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 26


7 Multiple group <strong>growth</strong> <strong>curves</strong><br />

Multiple group analysis <strong>using</strong> SEM is extremely flexible—some would say it is too flexible<br />

because there are so many possibilities. We use gender for our grouping variable because we are<br />

interested in the trend in BMI for girls compared to boys. We think of adolescent girls are more<br />

concerned about their weight <strong>and</strong> therefore more likely to have a lower BMI than boys <strong>and</strong> to have<br />

a flatter trajectory.<br />

There are several ways of comparing a model across multiple groups.<br />

One approach is to see if the same model fits each group, allowing all of the estimated parameters<br />

to be different.<br />

• Here we are saying that a linear <strong>growth</strong> model fits the data for both boys <strong>and</strong> girls, but<br />

• We are not constraining girls <strong>and</strong> boys to have the same values on any of the parameters. They<br />

may differ on the<br />

- intercept mean<br />

- slope mean<br />

- intercept variance<br />

- slope variance<br />

- covariance of intercept <strong>and</strong> slope residuals<br />

- residual errors<br />

- covariance of the residual errors that may be specified.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 27


• We can then put increasing invariance constraints on the model.<br />

a. At a minimum, we want to test whether the two groups have a different intercept (level) <strong>and</strong><br />

slope.<br />

b. If this constraint is acceptable we can add additional constraints on the variances,<br />

covariances, <strong>and</strong> error terms.<br />

First, we will estimate the model simultaneously for girls <strong>and</strong> boys with no constraints on the<br />

parameters. Here is the program with new comm<strong>and</strong>s highlighted:<br />

Title:<br />

bmi_<strong>growth</strong>_gender.inp<br />

Data:<br />

File is "C:\<strong>mplus</strong> examples\bmi_stata.dat" ;<br />

Variable:<br />

Names are<br />

id grlprb_y boyprb_y grlprb_p boyprb_p male<br />

race_eth bmi97 bmi98 bmi99 bmi00 bmi01 bmi02<br />

bmi03 white black hispanic asian other;<br />

Missing are all (-9999) ;<br />

! usevariables keeps bmi variables <strong>and</strong> gender<br />

Usevariables are male bmi97 bmi98 bmi99<br />

bmi00 bmi01 bmi02 bmi03 ;<br />

Grouping is male (0=female 1=male);<br />

Model:<br />

i s | bmi97@0 bmi98@1 bmi99@2 bmi00@3 bmi01@4<br />

bmi02@5 bmi03@6;<br />

Output:<br />

Sampstat Mod(3.84) ;<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00 bmi01 bmi02 bmi03(*);<br />

I’ve put the only changes we need to make in bold, underline.<br />

• We have a binary variable, male, that is coded 0 for females <strong>and</strong> 1 for males.<br />

• We add male to the list of variables we are <strong>using</strong>.<br />

• We add a subcomm<strong>and</strong> to the Variable: section that says we have a grouping variable,<br />

names it, <strong>and</strong> defines what the values are so the output will be labeled nicely.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 28


• The comm<strong>and</strong> Grouping is male (0=female 1 = male); is going to give us a<br />

separate set of estimates for the parameters for girls (labeled female) <strong>and</strong> boys (labeled<br />

male).<br />

• The estimation does both groups simultaneously.<br />

Here is selected, annotated output:<br />

SUMMARY OF ANALYSIS<br />

Number of groups 2<br />

Number of observations<br />

Group FEMALE 859<br />

Group MALE 909<br />

The following shows that we have the same variables in the model<br />

Number of dependent variables 7<br />

Number of independent variables 0<br />

Number of continuous latent variables 2<br />

Observed dependent variables<br />

Continuous<br />

BMI97 BMI98 BMI99 BMI00 BMI01 BMI02<br />

BMI03<br />

Continuous latent variables<br />

I S<br />

Variables with special functions<br />

Grouping variable MALE<br />

SAMPLE STATISTICS<br />

ESTIMATED SAMPLE STATISTICS FOR FEMALE<br />

Means<br />

BMI97 BMI98 BMI99 BMI00 BMI01<br />

________ ________ ________ ________ ________<br />

1 20.432 21.840 22.375 22.916 23.443<br />

Means<br />

BMI02 BMI03<br />

________ ________<br />

1 24.295 24.727<br />

ESTIMATED SAMPLE STATISTICS FOR MALE<br />

Means<br />

BMI97 BMI98 BMI99 BMI00 BMI01<br />

________ ________ ________ ________ ________<br />

1 20.698 21.848 22.896 23.665 24.220<br />

Means<br />

BMI02 BMI03<br />

________ ________<br />

1 24.467 25.111<br />

TESTS OF MODEL FIT<br />

Chi-Square Test of Model Fit<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 29


Value 411.966<br />

Degrees of Freedom 46 was 23<br />

P-Value 0.0000<br />

Chi-Square Contributions From Each Group<br />

FEMALE 150.775<br />

MALE 261.191<br />

Chi-Square Test of Model Fit for the Baseline Model<br />

Value 11735.530<br />

Degrees of Freedom 42<br />

P-Value 0.0000<br />

CFI/TLI<br />

CFI 0.969<br />

TLI 0.971<br />

Loglikelihood<br />

H0 Value -27639.607<br />

H1 Value -27433.624<br />

Information Criteria<br />

Number of Free Parameters 24<br />

Akaike (AIC) 55327.213<br />

Bayesian (BIC) 55458.676<br />

Sample-Size Adjusted BIC 55382.430<br />

(n* = (n + 2) / 24)<br />

RMSEA (Root Mean Square Error Of Approximation)<br />

Estimate 0.095<br />

90 Percent C.I. 0.087 0.103<br />

SRMR (St<strong>and</strong>ardized Root Mean Square Residual)<br />

Value 0.072<br />

MODEL RESULTS<br />

Two-Tailed<br />

Estimate S.E. Est./S.E. P-Value<br />

Group FEMALE<br />

S WITH<br />

I 0.522 0.103 5.050 0.000<br />

Means<br />

I 20.881 0.143 145.640 0.000<br />

S 0.663 0.025 27.015 0.000<br />

Variances<br />

I 15.141 0.859 17.626 0.000<br />

S 0.264 0.026 10.221 0.000<br />

Residual Variances<br />

BMI97 4.662 0.334 13.980 0.000<br />

BMI98 3.368 0.242 13.940 0.000<br />

BMI99 2.753 0.190 14.503 0.000<br />

BMI00 5.154 0.308 16.718 0.000<br />

BMI01 3.084 0.226 13.649 0.000<br />

BMI02 13.344 0.769 17.360 0.000<br />

BMI03 6.105 0.517 11.812 0.000<br />

Group MALE<br />

S WITH<br />

I 0.278 0.102 2.719 0.007<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 30


Means<br />

I 21.180 0.139 152.166 0.000<br />

S 0.732 0.024 30.661 0.000<br />

Variances<br />

I 14.911 0.824 18.094 0.000<br />

S 0.254 0.025 10.292 0.000<br />

Residual Variances<br />

BMI97 6.693 0.417 16.066 0.000<br />

BMI98 3.237 0.227 14.279 0.000<br />

BMI99 3.671 0.223 16.487 0.000<br />

BMI00 3.730 0.224 16.656 0.000<br />

BMI01 2.489 0.185 13.434 0.000<br />

BMI02 5.416 0.357 15.190 0.000<br />

BMI03 10.857 0.676 16.063 0.000<br />

Here is the graph of the two <strong>growth</strong> <strong>curves</strong>. It appears that the girls have a lower initial level <strong>and</strong> a<br />

flatter rate of <strong>growth</strong> of BMI.<br />

We should not rely on our visual inspection, but should explicitly test whether the girls <strong>and</strong> boys<br />

have a significant difference in their intercept <strong>and</strong> their slope. We can re-estimate the model with<br />

the intercept <strong>and</strong> slope invariant (or do it twice so we could have separate tests.) To do this we<br />

make the following modifications to the model:<br />

Notice that we added two lines to the Model: section,<br />

Title: bmi_<strong>growth</strong>_gender_equal.inp<br />

Data:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 31


Variable:<br />

Model:<br />

File is "C:\<strong>mplus</strong> examples\bmi_stata.dat" ;<br />

Names are<br />

id grlprb_y boyprb_y grlprb_p boyprb_p male<br />

race_eth bmi97 bmi98 bmi99 bmi00 bmi01 bmi02<br />

bmi03 white black hispanic asian other;<br />

Missing are all (-9999) ;<br />

Usevariables are male bmi97 bmi98 bmi99<br />

bmi00 bmi01 bmi02 bmi03 ;<br />

Grouping is male (0=female 1=male);<br />

i s | bmi97@0 bmi98@1 bmi99@2 bmi00@3 bmi01@4<br />

bmi02@5 bmi03@6;<br />

Model:<br />

[i] (1);<br />

[s] (2);<br />

Model male:<br />

[i] (1);<br />

[s] (2);<br />

Output: Sampstat Mod(3.84) ;<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00 bmi01 bmi02<br />

bmi03(*);<br />

We kept the group subcomm<strong>and</strong> <strong>and</strong> added the grouping variable, male. We added two Model:<br />

subcomm<strong>and</strong>s under the Model: comm<strong>and</strong>. The first one refers to the first group.<br />

• Since girls were coded 0 <strong>and</strong> boys were coded 1, the first group is girls. We put the name of<br />

parameters in square brackets, [i] <strong>and</strong> [s].<br />

• If we had called these initial <strong>and</strong> trend we would have typed [initial] <strong>and</strong> [trend].<br />

We put an arbitrary number in parentheses after the parameter name. Thus, we put (1)<br />

after [i].<br />

In the second subcomm<strong>and</strong>, Model male:, we put the name of the parameters followed by the<br />

same numbers as they had in the first gorup. Thus, the intercept gets the number 1 for girls <strong>and</strong><br />

also gets the number 1 for boys. This tells Mplus these must be held equal. They are still<br />

optimized, but with the constraint that they are equal.<br />

• If we had typed [i] (2) under Model male: what would happen? We would have<br />

constrained the boys intercept to be equal to the girls slope—not something we would want to<br />

do.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 32


• If we had omitted the [s] (2) under Model male: What would happen. We would have<br />

constrained both solutions to have the same intercept [i] (1), but not constrained them to<br />

have the same slopes.<br />

• The first Model: comm<strong>and</strong> is understood to be the group coded as zero on the male variable.<br />

• These changes force the intercept to be equal in both groups because they are both assigned<br />

parameter (1) <strong>and</strong> the slopes to be equal because they are both assigned a parameter (2).<br />

• Any parameters with a (1) after them are equal in both groups as are any parameters with a<br />

(2) after them in both groups.<br />

• Notice that we have square brackets [ ] around the names of the intercept <strong>and</strong> slope.<br />

When we run the revised program we obtain a chi-square that has two extra degrees of freedom<br />

because of the two constraints.<br />

TESTS OF MODEL FIT<br />

Chi-Square Test of Model Fit<br />

Value 418.884<br />

Degrees of Freedom 48<br />

P-Value 0.0000<br />

We had a chi-square(46) = 411.966 without these constraints. The<br />

difference has a chi-square of 6.918 with 2 degrees of freedom.<br />

Using Stata the significance is chi-square(2) = 6.918, p < .05<br />

. display 1-chi2(2,6.918)<br />

.03146121<br />

Chi-Square Contributions From Each Group<br />

The model fits females much better than it fits males:<br />

FEMALE 154.530<br />

MALE 264.353<br />

Chi-Square Test of Model Fit for the Baseline Model<br />

Value 11735.530<br />

Degrees of Freedom 42<br />

P-Value 0.0000<br />

CFI/TLI<br />

CFI 0.968<br />

TLI 0.972<br />

Loglikelihood<br />

H0 Value -27643.065<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 33


H1 Value -27433.624<br />

Information Criteria<br />

Number of Free Parameters 22<br />

Akaike (AIC) 55330.131<br />

Bayesian (BIC) 55450.638<br />

Sample-Size Adjusted BIC 55380.746<br />

(n* = (n + 2) / 24)<br />

RMSEA (Root Mean Square Error Of Approximation)<br />

Estimate 0.093<br />

90 Percent C.I. 0.085 0.102<br />

SRMR (St<strong>and</strong>ardized Root Mean Square Residual)<br />

Value 0.079<br />

MODEL RESULTS<br />

Two-Tailed<br />

Estimate S.E. Est./S.E. P-Value<br />

Group FEMALE<br />

S WITH<br />

I 0.528 0.104 5.097 0.000<br />

Means<br />

I 21.046 0.100 210.324 0.000<br />

S 0.700 0.017 40.814 0.000<br />

Group MALE<br />

S WITH<br />

I 0.281 0.102 2.742 0.006<br />

Means<br />

I 21.046 0.100 210.324 0.000<br />

S 0.700 0.017 40.814 0.000<br />

We should also look at the variances <strong>and</strong> covariances.<br />

• Although we can say there is a highly significant difference between the level <strong>and</strong> trend for<br />

girls <strong>and</strong> boys, we need to be cautious because this difference of chi-square has the same<br />

problem with a large sample size that the original chi-squares have.<br />

• In fact, the measures of fit are hardly changed whether we constrain the intercept <strong>and</strong> slope to<br />

be equal or not. Moreover, the visual difference in the graph is not dramatic.<br />

We could also put other constraints on the two solutions such as equal variances <strong>and</strong> covariances,<br />

<strong>and</strong> even equal residual error variances, but we will not.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 34


8 An Alternative to Multiple Group Analysis<br />

An alternative way of doing this, where there are two groups, is to enter the grouping variable as a<br />

predictor. This requires re-conceptualizing our model. We can think of the indicator variable<br />

Male having a direct path to both the intercept <strong>and</strong> the slope. Because the indicator variable is<br />

coded as 1 for male <strong>and</strong> 0 for female,<br />

• If the path from Male to the Intercept is positive this means that boys have a higher initial<br />

level on BMI.<br />

• Similarly, if there is a positive path from Male to the Slope, this indicates that boys have a<br />

steeper slope than girls on BMI. This direct effect actually represents an interaction between<br />

the trajectory <strong>and</strong> gender.<br />

• Such results would be consistent with our expectation that boys both start higher <strong>and</strong> gain more<br />

fat than girls during adolescence.<br />

• This approach does not let us test for other types of invariances such as the residual variances,<br />

covariances, <strong>and</strong> error terms.<br />

a. We are forcing these to be the same for both females <strong>and</strong> males; this may be unreasonable.<br />

b. The r<strong>and</strong>om effect for the slope for boys, R2, may be greater or less than it is for girls. We<br />

will not be able to evaluate this possibility with this approach.<br />

The following figure shows these two paths. We are explaining why some people have a higher or<br />

lower initial level <strong>and</strong> why some have a steeper or flatter slope by whether they are a girl or a boy.<br />

We are predicting that boys have a higher initial level <strong>and</strong> a steeper slope.<br />

Here is the figure:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 35


Here is the program:<br />

Title: bmi_gender_alternatives.inp<br />

bmi <strong>growth</strong> curve <strong>using</strong> gender as a single covariate.<br />

This is an alternative to <strong>using</strong> gender as two groups.<br />

Data: File is "c:\Mplus examples\bmi_stata.dat" ;<br />

Variable: Names are<br />

id grlprb_y boyprb_y grlprb_p boyprb_p male<br />

race_eth bmi97 bmi98 bmi99 bmi00 bmi01 bmi02 bmi03<br />

white black Hispanic asian other;<br />

Missing are all (-9999) ;<br />

Usevariables are male bmi97 bmi98 bmi99<br />

bmi00 bmi01 bmi02 bmi03 ;<br />

Model:<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 36


i s | bmi97@0 bmi98@1 bmi99@2 bmi00@3 bmi01@4 bmi02@5 bmi03@6;<br />

i on male;<br />

s on male;<br />

Output: Sampstat Mod(3.84) st<strong>and</strong>ardized;<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00 bmi01 bmi02<br />

bmi03(*);<br />

Here is selected, annotated output:<br />

SUMMARY OF ANALYSIS<br />

Number of groups<br />

1<br />

Number of observations<br />

1771<br />

TESTS OF MODEL FIT<br />

Chi-Square Test of Model Fit<br />

We cannot compare this chi-square to the two group chi-square<br />

because it is not a nested model.<br />

Value 301.244<br />

Degrees of Freedom 28<br />

P-Value 0.0000<br />

Chi-Square Test of Model Fit for the Baseline Model<br />

Value 11544.530<br />

Degrees of Freedom 28<br />

P-Value 0.0000<br />

CFI/TLI<br />

CFI 0.976<br />

TLI 0.976<br />

Loglikelihood<br />

H0 Value -29020.154<br />

H1 Value -28869.532<br />

Information Criteria<br />

Number of Free Parameters 14<br />

Akaike (AIC) 58068.308<br />

Bayesian (BIC) 58145.018<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 37


Sample-Size Adjusted BIC 58100.541<br />

(n* = (n + 2) / 24)<br />

RMSEA (Root Mean Square Error Of Approximation)<br />

Estimate 0.074<br />

90 Percent C.I. 0.067 0.082<br />

Probability RMSEA


BMI00 4.354 0.185 23.534 0.000<br />

BMI01 2.834 0.149 18.972 0.000<br />

BMI02 9.409 0.398 23.624 0.000<br />

BMI03 8.626 0.424 20.360 0.000<br />

I 15.045 0.597 25.198 0.000<br />

S 0.253 0.018 14.175 0.000<br />

R-SQUARE<br />

Observed Two-Tailed<br />

Variable Estimate S.E. Est./S.E. P-Value<br />

BMI97 0.724 0.012 59.339 0.000<br />

BMI98 0.832 0.009 93.431 0.000<br />

BMI99 0.846 0.008 110.551 0.000<br />

BMI00 0.820 0.008 99.975 0.000<br />

BMI01 0.888 0.006 138.194 0.000<br />

BMI02 0.731 0.011 65.511 0.000<br />

BMI03 0.772 0.011 70.133 0.000<br />

You can see why we rarely report the R-square for the intercept<br />

<strong>and</strong> slope.<br />

Latent Two-Tailed<br />

Variable Estimate S.E. Est./S.E. P-Value<br />

I 0.001 0.002 0.608 0.543<br />

S 0.007 0.006 1.266 0.206<br />

We see that the intercept is 20.385 <strong>and</strong> the slope is .625. How is gender related to this?<br />

For girls the equation is:<br />

Est. BMI = 20.911 + .656(Time) + .242(Male) + .086(Male)(Time)<br />

20.911 + .656(Time) + .242(0) + .086(0)(Time)<br />

= 20.911 + .656(Time)<br />

For boys the equation is:<br />

Est BMI = 20.911 + .656(Time) + .242(1) + .086(1)(Time)<br />

= (20.911 + .242) + (.625 + .086)(Time)<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 39


= 21.153 + .711(Time)<br />

Where Time is coded as 0, 1, 2, 3, 4, 5, 6<br />

Using these we estimate the BMI for girls is initially 20.911. By the seventh year when she is<br />

18(Time = 6) her estimated BMI will be 20.385 + .656(6) or 24.847<br />

Using these results, we estimate the BMI for boys is initially 21.153. By the seventh year it will be<br />

21.153 + .711(6) or 25.419. Since a BMI of 25 is considered overweight, by the age of 18 we<br />

estimate the average boy will be classified as overweight <strong>and</strong> the average girl is not far behind!<br />

We could use the plots provided by Mplus, but if we wanted a nicer looking plot we could use<br />

another program. I used Stata getting this graph.<br />

The Stata comm<strong>and</strong> is (this is driven by a drop down menu)<br />

twoway (connected Girls Age, lcolor(black) lpattern(dash) ///<br />

lwidth(medthick)) (connected Boys Age, lcolor(black) ///<br />

lpattern(solid) lwidth(medthick)), ///<br />

ytitle(Body Mass Index) xtitle(Age of Adolescent) ///<br />

caption(NLSY97 Data)<br />

<strong>and</strong> the data is<br />

+-----------------------+<br />

| Age Girls Boys |<br />

|-----------------------|<br />

1. | 12 20.911 21.153 |<br />

2. | 18 24.847 25.419 |<br />

+-----------------------+<br />

Body Mass Index for Adolescents<br />

Comparison of Girls <strong>and</strong> Boys<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 40


Limitations of this approach<br />

• When we treat a categorical variable as a grouping variable <strong>and</strong> do multiple comparisons we<br />

can test the equality of all the parameters.<br />

• When we treat it as a predictor as in this example, we only test whether the intercept <strong>and</strong><br />

slope are different for the two groups (interaction). In this example we do not allow the<br />

other parameters to be different for boys <strong>and</strong> girls <strong>and</strong> this might be a problem in some<br />

applications.<br />

9 Growth Curves with Time Invariant Covariates<br />

An extension of having a single categorical predictor includes having a series of covariates that<br />

explain variance in the intercept <strong>and</strong> slope. In this example we use what are known as time<br />

invariant covariates. These are covariates that either remain constant (gender) or for which you<br />

have a measure only at the start of the study. These are some times considered fixed effects since<br />

their value cannot change from one wave to another. It is possible to add time varying covariates<br />

as well.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 41


This has been called the Conditional Latent Trajectory Modeling (Curran & Hussong, 2003)<br />

because your initial level <strong>and</strong> trajectory (slope) are conditional on other variables.<br />

*The variable White (whites = 1; nonwhites = 0) compares Whites to the combination of African<br />

American <strong>and</strong> Hispanic. Asian & Pacific Isl<strong>and</strong>er, <strong>and</strong> Other have been deleted from this analysis<br />

because of small sample size.<br />

In this figure we have two covariates.<br />

• One is whether the adolescent<br />

is white versus African<br />

American or Hispanic <strong>and</strong><br />

• The other is a latent variable<br />

reflecting the level of<br />

emotional problems a youth<br />

has. There are two indicators<br />

of emotional problems, one<br />

from a parent report,<br />

boyprb_p, <strong>and</strong> the other from<br />

a youth report, boyprb_y.<br />

• The emotional problems are<br />

problems as reported at age 12.<br />

• A researcher may predict that<br />

Whites have a lower initial BMI<br />

(intercept) which persists during<br />

adolescence, but the White<br />

advantage does not increase (same<br />

slope as nonwhites).<br />

• Alternatively, a researcher may<br />

predict that being White predicts a<br />

lower initial BMI (intercept) <strong>and</strong><br />

less increase of the BMI (smaller<br />

slope) during adolescence.<br />

a. This suggests that minorities start with a disadvantage (high BMI) <strong>and</strong><br />

b. This disadvantaged gets even greater across adolescence.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 42


• A researcher may argue that emotional problems are associated with both higher initial BMI<br />

(intercept) <strong>and</strong> a more rapid increase in BMI over time (slope).<br />

• By including a covariate that is a latent variable itself, emotional problems, we will show how<br />

these are h<strong>and</strong>led by Mplus.<br />

We estimated this model for boys only; girls were excluded.<br />

The following is our Mplus program:<br />

Title: bmi_timea.inp<br />

bmi <strong>growth</strong> curve <strong>using</strong> race/ethnicity <strong>and</strong> emotional<br />

problems as a second covariate. There are two indicators<br />

of emotional problems.<br />

Data: File is "c:\Mplus examples\bmi_stata.dat" ;<br />

Variable: Names are<br />

id grlprb_y boyprb_y grlprb_p boyprb_p male race_eth bmi97<br />

bmi98 bmi99 bmi00 bmi01 bmi02 bmi03 white black hispanic<br />

asian other;<br />

Missing are all (-9999) ;<br />

Usevariables are boyprb_y boyprb_p white bmi97 bmi98 bmi99<br />

bmi00 bmi01 bmi02 bmi03 ;<br />

Useobservations = male eq 1 <strong>and</strong> asian ne 1 <strong>and</strong> other ne 1;<br />

Model: i s q| bmi97@0 bmi98@1 bmi99@2 bmi00@3 bmi01@4 bmi02@5<br />

bmi03@6;<br />

emot_prb by boyprb_p boyprb_y ;<br />

i on white emot_prb;<br />

s on white emot_prb;<br />

q on white emot_prb;<br />

Output: Sampstat Mod(3.84) st<strong>and</strong>ardized;<br />

Plot:<br />

Type is Plot3;<br />

Series = bmi97 bmi98 bmi99 bmi00 bmi01 bmi02 bmi03(*);<br />

I have highlighted the new lines in the Mplus program.<br />

• The format of the Useobservations subcomm<strong>and</strong> is similar to if or select used<br />

with other programs.<br />

• The Useobservations = male eq 1 <strong>and</strong> asian ne 1 <strong>and</strong> other ne 1;<br />

restricts our sample to males (male eq 1). This is very h<strong>and</strong>y when <strong>using</strong> the same<br />

dataset for a variety of models where you want some models to only include selected<br />

participants.<br />

• We have dropped Asians <strong>and</strong> members of the “other” category. There are relatively few of<br />

them in this sample dataset <strong>and</strong> they may have very different BMI trajectories. Also, the<br />

meaning of the category “other” is ambiguous.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 43


• I added a quadratic term in the Model: comm<strong>and</strong>. I first estimated this model <strong>using</strong> just a<br />

linear slope <strong>and</strong> the fit was not very good (results not shown here). Adding the quadratic<br />

improved the fit.<br />

• This example has a measurement model for a latent covariate, emot_prb. In other<br />

programs this can involve complicated programming. Here it is done with the single line.<br />

(You usually would like to have 3 <strong>and</strong> preferably 4 indicators of a latent variable.)<br />

emot_prb by boyprb_p boyprb_y ;<br />

• The by is a key word in Mplus for creating latent variables used in Confirmatory Factor<br />

Analysis <strong>and</strong> SEM.<br />

• On the right of the by are two observed variables. The boyprb_p is the report of parents<br />

about the adolescent’s emotional problems. The boyprb_y is the youths own report.<br />

• It is desirable to have three or more indicators of a latent variable, but we only have two<br />

here so that will have to do.<br />

• To the left of the by is the name we give to the latent variable, emot_prb. This new latent<br />

variable did not appear in the list of variables we are <strong>using</strong>, but it is defined here.<br />

• The “by” term<br />

o fixes the first variable to the right as a reference indicator, boyprb_p, <strong>and</strong> assigns a<br />

loading of 1 to it.<br />

o It lets the loading of the second variable, boyprb_y, be estimated. It also creates<br />

error/residual variances that are labeled e1 <strong>and</strong> e2 in the figure.<br />

o The default is that these errors are uncorrelated.<br />

o It is good practice to have the strongest indicator on the right of the “by” be the<br />

reference indicator with a loading fixed at 1.0. You can run the model <strong>and</strong> if this does<br />

not happen, you can re-run it, reversing the order of the items on the right of the<br />

“by.”<br />

• The next three new lines,<br />

o i on white emot_prb;<br />

o s on white emot_prb; <strong>and</strong><br />

o q on white emot_prb;<br />

o Define the relationship between the covariates <strong>and</strong> the intercept <strong>and</strong> slope. These<br />

represent interactions of each covariate with the intercept <strong>and</strong> slope.<br />

o These are the γ1wi in the equation for HLM users.<br />

o Mplus uses the on comm<strong>and</strong> to signify that a variable depends on another variable in<br />

the structural part of the model. The by comm<strong>and</strong> is the key to underst<strong>and</strong>ing how<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 44


Mplus sets up the measurement model <strong>and</strong> the on is the key to how Mplus sets up the<br />

structural model.<br />

There are many defaults. Mplus assumes there are residual variances <strong>and</strong> covariances for the<br />

intercept <strong>and</strong> slopes. It fixes the intercepts at zero. It assumes the intercept <strong>and</strong> slope variances are<br />

correlated.<br />

Here is selected results:<br />

Mplus VERSION 5.1<br />

MUTHEN & MUTHEN<br />

07/01/2008 8:01 PM<br />

TESTS OF MODEL FIT<br />

Chi-Square Test of Model Fit<br />

Value 88.824<br />

Degrees of Freedom 34<br />

P-Value 0.0000<br />

Chi-Square Test of Model Fit for the Baseline Model<br />

Value 5975.020<br />

Degrees of Freedom 45<br />

P-Value 0.0000<br />

CFI/TLI<br />

CFI 0.991<br />

TLI 0.988<br />

Loglikelihood<br />

H0 Value -17221.918<br />

H1 Value -17177.507<br />

Information Criteria<br />

Number of Free Parameters 29<br />

Akaike (AIC) 34501.837<br />

Bayesian (BIC) 34640.190<br />

Sample-Size Adjusted BIC 34548.093<br />

(n* = (n + 2) / 24)<br />

RMSEA (Root Mean Square Error Of Approximation)<br />

Estimate 0.043<br />

90 Percent C.I. 0.032 0.054<br />

Probability RMSEA


Measurement of latent variable.<br />

EMOT_PRB BY<br />

BOYPRB_P 1.000 0.000 999.000 999.000<br />

BOYPRB_Y 0.575 0.171 3.374 0.001<br />

Emotional problems does not have a significant effect on the initial level<br />

at age 12, but significantly increases the slope. Significant negative<br />

effect on quadratic is a bit conf<strong>using</strong>.<br />

I ON<br />

EMOT_PRB 0.300 0.168 1.793 0.073<br />

S ON<br />

EMOT_PRB 0.212 0.089 2.370 0.018<br />

Q ON<br />

EMOT_PRB -0.037 0.015 -2.462 0.014<br />

Whites have a significant advantage initially (intercept), but there is not<br />

a significant compounding of this over time since White does not<br />

significantly influence the slope or quadratic.<br />

I ON<br />

WHITE -1.030 0.292 -3.529 0.000<br />

S ON<br />

WHITE 0.130 0.138 0.941 0.346<br />

Q ON<br />

WHITE -0.030 0.023 -1.293 0.196<br />

S WITH<br />

I 0.701 0.329 2.128 0.033<br />

Q WITH<br />

I -0.101 0.053 -1.922 0.055<br />

S -0.174 0.034 -5.081 0.000<br />

WHITE WITH<br />

EMOT_PRB -0.111 0.028 -4.013 0.000<br />

Intercepts<br />

BOYPRB_Y 2.108 0.052 40.712 0.000<br />

BOYPRB_P 1.893 0.058 32.668 0.000<br />

BMI97 0.000 0.000 999.000 999.000<br />

BMI98 0.000 0.000 999.000 999.000<br />

BMI99 0.000 0.000 999.000 999.000<br />

BMI00 0.000 0.000 999.000 999.000<br />

BMI01 0.000 0.000 999.000 999.000<br />

BMI02 0.000 0.000 999.000 999.000<br />

BMI03 0.000 0.000 999.000 999.000<br />

I 21.279 0.210 101.495 0.000<br />

S 1.171 0.100 11.719 0.000<br />

Q -0.077 0.016 -4.649 0.000<br />

The linear slope of 1.171 is huge when you project this over the six years.<br />

The quadratic slope being negative indicates that there is some leveling off<br />

in the increase in BMI.<br />

Variances<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 46


EMOT_PRB 1.485 0.455 3.262 0.001<br />

Residual Variances<br />

BOYPRB_Y 1.850 0.170 10.875 0.000<br />

BOYPRB_P 1.203 0.442 2.722 0.006<br />

BMI97 5.535 0.479 11.564 0.000<br />

BMI98 3.276 0.227 14.402 0.000<br />

BMI99 3.333 0.221 15.067 0.000<br />

BMI00 3.182 0.218 14.586 0.000<br />

BMI01 2.361 0.186 12.670 0.000<br />

BMI02 5.225 0.357 14.654 0.000<br />

BMI03 8.961 0.743 12.057 0.000<br />

I 13.094 0.869 15.070 0.000<br />

S 1.291 0.215 6.007 0.000<br />

Q 0.030 0.006 5.092 0.000<br />

STANDARDIZED MODEL RESULTS<br />

STDYX St<strong>and</strong>ardization<br />

Two-Tailed<br />

Estimate S.E. Est./S.E. P-Value<br />

EMOT_PRB BY<br />

BOYPRB_P 0.743 0.111 6.708 0.000<br />

BOYPRB_Y 0.458 0.073 6.308 0.000<br />

Notice the z-tests are slightly different. Most st<strong>and</strong>ard packages assume the<br />

unst<strong>and</strong>ardized test is the same.<br />

I ON<br />

EMOT_PRB 0.099 0.052 1.909 0.056<br />

S ON<br />

EMOT_PRB 0.222 0.080 2.775 0.006<br />

Q ON<br />

EMOT_PRB -0.253 0.087 -2.918 0.004<br />

I ON<br />

WHITE -0.140 0.039 -3.558 0.000<br />

S ON<br />

WHITE 0.056 0.059 0.942 0.346<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 47


Q ON<br />

WHITE -0.083 0.064 -1.293 0.196<br />

S WITH<br />

I 0.170 0.090 1.891 0.059<br />

Q WITH<br />

I -0.163 0.094 -1.731 0.083<br />

S -0.891 0.021 -42.268 0.000<br />

WHITE WITH<br />

EMOT_PRB -0.183 0.049 -3.748 0.000<br />

Residual Variances<br />

BOYPRB_Y 0.790 0.067 11.864 0.000<br />

BOYPRB_P 0.448 0.165 2.717 0.007<br />

BMI97 0.290 0.025 11.460 0.000<br />

BMI98 0.171 0.012 13.889 0.000<br />

BMI99 0.151 0.011 13.794 0.000<br />

BMI00 0.132 0.010 13.364 0.000<br />

BMI01 0.096 0.008 11.571 0.000<br />

BMI02 0.185 0.013 14.176 0.000<br />

BMI03 0.269 0.022 12.260 0.000<br />

I 0.966 0.016 62.173 0.000<br />

S 0.952 0.034 27.829 0.000<br />

Q 0.937 0.043 22.016 0.000<br />

Unfortunately, we cannot get graphs when we have covariates. You could create these yourself by<br />

substituting fix values for race <strong>and</strong> emotional problems.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 48


10 Mediational Models with Time Invariant Covariates<br />

Sometimes all of the covariates are time invariant or at least measured at just the start of the study.<br />

Curran <strong>and</strong> Hussong (2003) discuss a study of a latent <strong>growth</strong> curve on drinking problems with a<br />

covariate of parental drinking. Parental drinking influences both the initial level <strong>and</strong> the rate of<br />

<strong>growth</strong> of drinking problem behavior among adolescents. The question is whether some other<br />

variables might mediate this relationship<br />

• Parental monitoring<br />

• Peer influence<br />

Mplus allows us to estimate the direct <strong>and</strong> indirect effect of Parent Drinking on the Intercept <strong>and</strong><br />

Slope. It also provides a test of significance for these effects.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 49


11 Time Varying Covariates<br />

We have illustrated time invariant covariates that are measured at time 1. It is possible to extend<br />

this to include time varying covariates. Time varying covariates either are measured after the<br />

process has started or have a value that changes (hours of nutrition education, level of program<br />

fidelity). Although we will not show our output, we will illustrate the use of time varying<br />

covariates in a figure. In this figure the time varying covariates, a21 to a24 might be<br />

• Hours of nutrition education completed between waves. Independent of the overall <strong>growth</strong><br />

trajectory, η1, students who have several hours of nutrition education programming may have a<br />

decrease in their BMI<br />

• Physical education curriculum. A physical activity program might lead to reduced BMI.<br />

Students who spend more time in this physical activity program might have a lower BMI<br />

independent of the overall <strong>growth</strong> trend. Hours in physical education courses will vary from<br />

year to year.<br />

• This would be a good way to incorporate fidelity into a program evaluation.<br />

This figure is borrowed from Muthén where he is examining <strong>growth</strong> in math performance over 4<br />

years. The w vector contains x variables are covariates that directly influence the intercept, η0, or<br />

slope, η1. The aij are number of math courses taken each year.<br />

yit = repeated measures on the outcome (math achievement)<br />

a1it = Time score (0, 1, 2, 3) as discussed previously<br />

a2it = Time varying covariates<br />

(# of math courses taken that year)<br />

w = Vector of x covariates that<br />

are time invariant <strong>and</strong> measured at<br />

or before the first yit<br />

In this example we might think of<br />

the yi variables being measures of<br />

conflict behavior where y1 is at age<br />

17 <strong>and</strong> y4 is at age 25. We know<br />

there is a general decline in<br />

conflict behavior during this time<br />

interval. Therefore, the slope η1 is<br />

expected to be negative.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 50


Now suppose we also have a measure of alcohol abuse for each of the 4 waves (aij). We might<br />

hypothesize that during a year in which an adolescent has a high score on alcohol abuse (say<br />

number of days the person drinks 5 or more drinks in the last 30 days) that there will be an<br />

elevated level of conflict behavior that cannot be explained by the general decline (negative<br />

slope).<br />

The negative slope reflects the general decline in conflict behavior by young adults as the move<br />

from age 17 to age 25. The effect of aij on yi provides the additional explanation that those years<br />

when there is a lot of drinking; there will be an elevated level of conflict that does not fit the<br />

general decline.<br />

If you want more, here are a few references<br />

2. Basic <strong>growth</strong> curve modeling<br />

a. Bollen, K. A., & Curran, P. J. (2006). Latent Curve Models: A Structural Equation<br />

Perspective. Hoboken, NJ: Wiley.<br />

b. Curran, F. J., & Hussong, A. M. (2003). The Use of latent Trajectory Models in<br />

Psychopathology Research. Journal of Abnormal Psychology. 112:526-544. This is a<br />

general introduction to <strong>growth</strong> <strong>curves</strong> that is accessible.<br />

c. Duncan, T. E., Duncan, S. C., & Strycker, L. A. (2006). An Introduction to Latent<br />

Variable Growth Curve Modeling: Concepts, Issues, <strong>and</strong> Applications (2 nd ed.).<br />

Mahwah NJ: Lawrence Erlbaum. The second edition of a classic text on <strong>growth</strong> curve<br />

modeling.<br />

d. Kaplan, D. (2000). Chapter 8: Latent Growth Curve Modeling. In D. Kaplan,<br />

Structural Equation Modeling: Foundations <strong>and</strong> Extensions (pp 149-170). Thous<strong>and</strong><br />

Oaks, CA: Sage. This is a short overview.<br />

e. Wang, M. (2007). Profiling retirees in the retirement transition <strong>and</strong> adjustment<br />

process: Examining the longitudinal change patterns of retirees' psychological wellbeing.<br />

Journal of Applied Psychology, 92(2), 455-474. This is a nice example of<br />

presenting results showing some graphs <strong>and</strong> tables.<br />

3. Limited Outcome Variables: Binary <strong>and</strong> count variables<br />

a. Muthén, B. (1996). Growth modeling with binary responses. In A. V. Eye & C.<br />

Clogg (Eds.) Categorical Variables in Developmental Research: Methods of analysis<br />

(pp 37-54). San Diego, CA: Academic Press.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 51


. Long, J. S., & Freese, J. (2006). Regression Models for Categorical Dependent<br />

Variables Using Stata, 2 nd ed. Stata Press (www.stata-press.com). This provides the<br />

most accessible <strong>and</strong> still rigorous treatment of how to use an interpret limited<br />

dependent variables.<br />

c. Rabe-Hesketh, S., & Skrondal, A. (2005). Multilevel <strong>and</strong> Longitudinal Modeling<br />

Using Stata. Stata Press (www.stata-press.com). This discusses a free set of<br />

comm<strong>and</strong>s that can be added to Stata that will do most of what Mplus can do <strong>and</strong><br />

some things Mplus cannot do. It is hard to use <strong>and</strong> very slow.<br />

4. Growth mixture modeling<br />

a. Muthén, B., & Muthén, L. K. (2000). Integrating person-centered <strong>and</strong> variablecentered<br />

analysis: Growth mixture modeling with latent trajectory classes.<br />

Alcoholism: Clinical <strong>and</strong> Experimental Research. 24:882-891.<br />

This is an excellent <strong>and</strong> accessible conceptual introduction.<br />

b. Muthén, B. (2001). Latent variable mixture modeling. In G. Marcoulides, & R.<br />

Schumacker (Eds.) New Developments <strong>and</strong> Techniques in Structural Equation<br />

Modeling (pp. 1-34). Mahwah, NJ: Lawrence Erlbaum.<br />

c. Muthén, B., Brown, C. H., Booil, J., Khoo, S. Yang, C. Wang, C., Kellam, S., Carlin,<br />

J., & Liao, J. (2002). General <strong>growth</strong> mixture modeling for r<strong>and</strong>omized preventive<br />

interventions. Biostatistics, 3:459-475<br />

d. Muthén, B. Latent Variable analysis: Growth Mixture Modeling <strong>and</strong> Related<br />

Techniques for Longitudinal Data. (2004) In D. Kaplan (ed.), H<strong>and</strong>book of<br />

quantitative methodology for the social sciences (pp. 345-368). Newbury Park, CA:<br />

Sage Publications<br />

e. Muthén, B., Brown, C. H., Booil Jo, K, M., Khoo, S., Yang, C. Wang, C., Kellam, S.,<br />

Carlin, J., Liao, J. (2002). General <strong>growth</strong> mixture modeling for r<strong>and</strong>omized<br />

preventive interventions. Biostatistics. 3,4, pp. 459-475.<br />

5. The web page for Mplus, www.statmodel.com , maintains a current set of references, many<br />

as PDF files. These are organized by topic <strong>and</strong> some include data <strong>and</strong> the Mplus program.<br />

Growth Curve <strong>and</strong> Related Models, Alan C. Acock 52

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!