-
Notifications
You must be signed in to change notification settings - Fork 0
/
lpa2.rmd
349 lines (269 loc) · 10.8 KB
/
lpa2.rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
---
title: "Latent Profile Analysis using Mclust"
output:
html_document:
toc: true
toc_float: true
collapsed: false
number_sections: false
toc_depth: 1
code_folding: show
code_download: yes
---
```{r setup, include=FALSE}
knitr::opts_chunk$set(message=FALSE,warning=FALSE, cache=TRUE)
```
<br>
## PART 2
*Last edited: January 20, 2021*
<br>
Let's now estimate the LPA model with as input the CFA factor-scores. Again, the idea is that this would provide subscale estimates that consider the unique contribution of each item (instead of averaging over all items), resulting in "true" motivational configurations (see [this article](https://scholarworks.umass.edu/cgi/viewcontent.cgi?article=1226&context=pare)).
We are going to use the [*Lavaan*](http://www.understandingdata.net/2017/03/22/cfa-in-lavaan/)-package for CFA.
---
# 1. Data
Let's call our 'cleaned' SMS-data again.
```{r, warning=FALSE, message=FALSE}
library(sjlabelled)
library(dplyr)
library(tidyverse)
library(careless)
library(psych)
load("data_abs_public_v2.RData") # load data
data_abs_public <- unlabel(data_abs_public, verbose=F)
sms <- data_abs_public %>% select(W1_M1_1, W1_M1_2, W1_M1_3, W1_M1_4, W1_M1_5, W1_M1_6, W1_M1_7, W1_M1_8, W1_M2_1, W1_M2_2, W1_M2_3, W1_M2_4, W1_M2_5, W1_M2_6, W1_M2_7, W1_M2_8, W1_M3_1, W1_M3_2, W1_M3_3, W1_M3_4, W1_M3_5, W1_M3_6, W1_M3_7, W1_M3_8) # subset W1
sms$id <- 1:length(sms[, 1]) # add identifier
sms <- sms %>%
mutate(string = longstring(.)) %>%
mutate(md = outlier(., plot = FALSE)) # make string variable
cutoff <- (qchisq(p = 1 - .001, df = ncol(sms)))
sms_clean <- sms %>%
filter(string <= 10,
md < cutoff) %>%
select(-string, -md) # cap string responding and use MD
```
<br>
----
# 2. CFA
Now we are going to perform confirmatory factor analysis using Lavaan. First, tell Lavaan the confirmatory structure.
```{r}
library(lavaan)
motivation_model <- "
amotivation =~ W1_M1_5 + W1_M2_4 + W1_M3_1 + W1_M3_6
external =~ W1_M1_4 + W1_M2_3 + W1_M3_3 + W1_M3_8
introjected =~ W1_M1_7 + W1_M2_2 + W1_M2_8 + W1_M3_7
identified =~ W1_M1_3 + W1_M1_8 + W1_M2_7 + W1_M3_4
integrated =~ W1_M1_2 + W1_M2_1 + W1_M2_5 + W1_M3_5
intrinsic =~ W1_M1_1 + W1_M1_6 + W1_M2_6 + W1_M3_2
id ~~ id"
```
When performing CFA (or any latent variable modeling) the question remains what scale to assign to the latent variable. Many statistical programs (including Lavaan) fix the loading of the first variable for a given latent variable to 1, which assigns the scale of that manifest indicator to the latent variable. We will use a standardized scale for the latent variable (i.e. mean = 0, sd = 1): the loading of the first variable is then freely estimated.
The default treatment of missing data is listwise deletion; perhaps not what we want. Later on we can try full information maximum likelihood (FIML), which will generally result in estimates similar to what you would get with multiple impution, but with the added advantage that it's all done in one step.
```{r echo=T, results='hide'}
fit <- cfa(motivation_model, data=sms_clean, # use the 'clean' data
std.lv=T # this tells lavaan to use a standardized scale for the latent variables.
)
```
Let's call the CFA output with the *knitr*-package.
```{r}
library(knitr)
options(knitr.kable.NA = '') # this will hide missing values in the kable table
parameterEstimates(fit, standardized=TRUE) %>%
filter(op == "=~") %>%
mutate(stars = ifelse(pvalue < .001, "***", ifelse(pvalue < .01, "**", ifelse(pvalue < .05, "*", "")))) %>%
select('Latent Factor'=lhs,
Indicator=rhs,
B=est,
SE=se, Z=z,
Beta=std.all,
sig=stars) %>%
kable(digits = 3, format="pandoc", caption="Table 1. Latent Factor Loadings"
)
```
So, as an example, for a 1-unit increase in the latent variable *amotivation* (i.e. 1-SD increase), the model predicts a 0.367 increase in W1_M1_5.
<br>
Let's also get the correlations between the factors.
```{r}
parameterEstimates(fit, standardized=TRUE) %>%
filter(op == "~~",
lhs %in% c("amotivation", "external", "introjected", "identified", "integrated", "intrinsic"),
!is.na(pvalue)) %>%
mutate(stars = ifelse(pvalue < .001, "***",
ifelse(pvalue < .01, "**",
ifelse(pvalue < .05, "*", "")))) %>%
select('Factor 1'=lhs,
'Factor 2'=rhs,
Correlation=est,
sig=stars) %>%
kable(digits = 3, format="pandoc", caption="Table 2: Latent Factor Correlations")
```
<br>
Now extract the factor scores, put them into a dataframe, delete missings, and standardize the factor scores (again). Also, we make a descriptive table.
```{r}
fs <- lavPredict(fit, newdata = sms_clean, append.data = TRUE) # extract
fs <- as.data.frame(fs[, c(1:6, 31)]) # subset factor scores and identifier in df
clus <- fs %>%
select(-id) %>%
na.omit() %>% # listwise deletion
mutate_all(list(scale)) # standardize
clus$id <- fs$id # add id again...
library(dplyr)
library(tidyr)
library(knitr)
library(kableExtra)
# describe
input <- clus %>%
select(-id) %>%
gather("Variable", "value") %>%
group_by(Variable) %>%
summarise(Mean=mean(value, na.rm=TRUE),
SD=sd(value, na.rm=TRUE),
min=min(value, na.rm=TRUE),
max=max(value, na.rm=TRUE))
knitr::kable(input, digits=2, "html", caption="Table 3. Descriptives of standardized factor scores SMS W1") %>%
kableExtra::kable_styling(bootstrap_options = c("striped", "hover"))
```
<br>
And check the distribution of our variables
```{r}
library(Hmisc)
hist(clus %>%
select(-id))
```
The produced standardized scores somewhat approximate a Z-score metric (i.e. where values range from ca. -3 to +3). Note that *amotivation* is an exception to the rule: it is very much right-skewed and has high kurtosis.
<br>
----
# 3. LPA
Let's follow the same procedure for LPA, but now we include the standardized factor scores as input.
## Model fit {.tabset .tabset-fade}
Starting off by exploring model fit again, by plotting it.
### BIC
```{r class.source = 'fold-hide'}
library(mclust)
BIC <- mclustBIC(clus %>%
select(-id)) # exclude id
plot(BIC)
```
### ICL
```{r class.source = 'fold-hide'}
library(mclust)
ICL <- mclustICL(clus %>%
select(-id))
plot(ICL)
```
### BLRT
```{r eval = FALSE}
library(mclust)
mclustBootstrapLRT(clus %>%
select(-id), modelName = "VEV")
```
## {-}
And use the *summary*-function to show the top-three models based on BIC and ICL.
## {.tabset .tabset-fade}
### BIC
```{r class.source = 'fold-hide'}
summary(BIC)
```
### ICL
```{r class.source = 'fold-hide'}
summary(ICL)
```
## {-}
<br>
----
# 4. Visualizing LPA
## {.tabset .tabset-fade}
Statistically, the best model is the VEV, 8. We do not plot this, but we estimate a sequence of models, incrementally increasing the number of profiles, until they do not provide theoretical additions to the model. Plotting the models will make it easier to, intuitively, assess which model is suitable. We start with 2 profiles.
### 2 profiles
```{r class.source = 'fold-hide'}
m2 <- Mclust(clus %>%
select(-id),
modelNames = "VEV", G = 2, x = BIC)
summary(m2)
means <- data.frame(m2$parameters$mean,
stringsAsFactors = F) %>%
rownames_to_column() %>%
rename(Motivation = rowname) %>%
melt(id.vars = "Motivation", variable.name = "Profile", value.name = "Mean") %>%
mutate(Mean = round(Mean, 2))
means %>%
ggplot(aes(Motivation, Mean, group = Profile, color = Profile)) +
geom_line() +
geom_point() +
scale_x_discrete(limits = c("amotivation", "external", "introjected", "identified", "integrated", "intrinsic")) +
labs(x = NULL, y = "Standardized mean latent SMS factor scores") +
theme_bw(base_size = 14) +
geom_hline(yintercept = 0, linetype="dashed") +
theme(axis.text.x = element_text(angle = 45, hjust = 1), legend.position = "top")
```
### 3 profiles
```{r class.source = 'fold-hide'}
m3 <- Mclust(clus %>%
select(-id),
modelNames = "VEV", G = 3, x = BIC)
summary(m3)
means <- data.frame(m3$parameters$mean,
stringsAsFactors = F) %>%
rownames_to_column() %>%
rename(Motivation = rowname) %>%
melt(id.vars = "Motivation", variable.name = "Profile", value.name = "Mean") %>%
mutate(Mean = round(Mean, 2))
means %>%
ggplot(aes(Motivation, Mean, group = Profile, color = Profile)) +
geom_line() +
geom_point() +
scale_x_discrete(limits = c("amotivation", "external", "introjected", "identified", "integrated", "intrinsic")) +
labs(x = NULL, y = "Standardized mean latent SMS factor scores") +
theme_bw(base_size = 14) +
geom_hline(yintercept = 0, linetype="dashed") +
theme(axis.text.x = element_text(angle = 45, hjust = 1), legend.position = "top")
```
### 4 profiles
```{r class.source = 'fold-hide'}
m4 <- Mclust(clus %>%
select(-id),
modelNames = "VEV", G = 4, x = BIC)
summary(m4)
means <- data.frame(m4$parameters$mean,
stringsAsFactors = F) %>%
rownames_to_column() %>%
rename(Motivation = rowname) %>%
melt(id.vars = "Motivation", variable.name = "Profile", value.name = "Mean") %>%
mutate(Mean = round(Mean, 2))
means %>%
ggplot(aes(Motivation, Mean, group = Profile, color = Profile)) +
geom_line() +
geom_point() +
scale_x_discrete(limits = c("amotivation", "external", "introjected", "identified", "integrated", "intrinsic")) +
labs(x = NULL, y = "Standardized mean latent SMS factor scores") +
theme_bw(base_size = 14) +
geom_hline(yintercept = 0, linetype="dashed") +
theme(axis.text.x = element_text(angle = 45, hjust = 1), legend.position = "top")
```
### 5 profiles
```{r class.source = 'fold-hide'}
m5 <- Mclust(clus %>%
select(-id),
modelNames = "VEV", G = 5, x = BIC)
summary(m5)
means <- data.frame(m5$parameters$mean,
stringsAsFactors = F) %>%
rownames_to_column() %>%
rename(Motivation = rowname) %>%
melt(id.vars = "Motivation", variable.name = "Profile", value.name = "Mean") %>%
mutate(Mean = round(Mean, 2))
means %>%
ggplot(aes(Motivation, Mean, group = Profile, color = Profile)) +
geom_line() +
geom_point() +
scale_x_discrete(limits = c("amotivation", "external", "introjected", "identified", "integrated", "intrinsic")) +
labs(x = NULL, y = "Standardized mean latent SMS factor scores") +
theme_bw(base_size = 14) +
geom_hline(yintercept = 0, linetype="dashed") +
theme(axis.text.x = element_text(angle = 45, hjust = 1), legend.position = "top")
```
## {-}
<br>
----
The output shows less theoretically interpretable motivational profiles compared to the previous LPA...
As a last approach (see next page), we will use weighted sum scores (as described on page 3 of [this article](https://scholarworks.umass.edu/cgi/viewcontent.cgi?article=1226&context=pare
)).