the power of fake data simulations
Posted onThis post is basically a self exercise of what Andrew Gelman has already posted here. Fake data simulations are incredible tools to understand your study. It forces you to think about what is the size of the effect you wish to see, what kind of variance is in your model, if you can really detect it, will your design work, and the list goes on. Similar to any practice, when you have to think critically and put things to paper you tend to see the weaknesses of your arguments. It also helos you to anticipate issues. All of these things are priceless.
So taking his work and looking at a between person and within person design let’s copy his code and build the fake data.
## 2. Simulate a data structure with N_per_person measurements on each of J people
J <- 50 # number of people in the experiment
N_per_person <- 10 # number of measurements per person
person_id <-
index <-
time <- index - 1 # time of measurements, from 0 to 9
N <-
a <-
b <-
theta <- 1
sigma_y <- 1
## 3. Simulate data from a between-person experiment
z <-
y_pred <- a + b*time + theta*z*time
y <-
z_full <- z
exposure <- z_full*time
data_1 <-
## 4. Simulate data from a within-person experiment: for each person, do one treatment for the first half of the experiment and the other treatment for the second half.
z_first_half <- z
T_switch <-
z_full <-
for (j in 1:J){
exposure <-
}
y_pred <- a + b*time + theta*exposure
y <-
data_2 <-
Just for clarity I am going to show a few records of the two different data sets:
time | person_id | exposure | y |
---|---|---|---|
0 | 1 | 0 | 1.571653 |
1 | 1 | 0 | 2.605884 |
2 | 1 | 0 | 2.684796 |
3 | 1 | 0 | 4.016867 |
4 | 1 | 0 | 6.582267 |
5 | 1 | 0 | 7.388390 |
Sample of Between Persons Data
data_2 %>%
%>%
knitr::
time | person_id | exposure | y |
---|---|---|---|
0 | 1 | 0 | -0.4361502 |
1 | 1 | 0 | 1.7448124 |
2 | 1 | 0 | 0.9436166 |
3 | 1 | 0 | 3.6318851 |
4 | 1 | 0 | 4.4409258 |
5 | 1 | 1 | 7.9628482 |
Sample of Within Persons Data
Now we can plot the data:
for (j in 1:J){
ok <- data_1$person_id==j
if (z == 0){
}
}
for (j in 1:J){
ok <- data_1$person_id==j
if (z == 1){
}
}
for (j in 1:J){
ok <- person_id==j
if (z == 0) {
}
}
for (j in 1:J){
ok <- person_id==j
if (z == 1) {
for (i in 1:N_per_person) {
ok2 <- ok & index==i
}
}
}
And then start the analysis using our HLM:
fit_1 <-
fit_2 <-
Between Persons Design Summary
stan_glmer
family: gaussian [identity]
formula: y ~ (1 + time | person_id) + time + exposure
observations: 500
------
Median MAD_SD
(Intercept) -0.1 0.2
time 0.9 0.2
exposure 1.3 0.3
Auxiliary parameter(s):
Median MAD_SD
sigma 1.0 0.0
Error terms:
Groups Name Std.Dev. Corr
person_id (Intercept) 1.1
time 1.1 -0.09
Residual 1.0
Num. levels: person_id 50
------
* For help interpreting the printed output see ?print.stanreg
* For info on the priors used see ?prior_summary.stanreg
Within Persons Design Summary
stan_glmer
family: gaussian [identity]
formula: y ~ (1 + time | person_id) + time + exposure
observations: 500
------
Median MAD_SD
(Intercept) 0.0 0.2
time 1.0 0.1
exposure 1.0 0.1
Auxiliary parameter(s):
Median MAD_SD
sigma 1.0 0.0
Error terms:
Groups Name Std.Dev. Corr
person_id (Intercept) 1.0
time 1.1 0.05
Residual 1.0
Num. levels: person_id 50
------
* For help interpreting the printed output see ?print.stanreg
* For info on the priors used see ?prior_summary.stanreg
The big take away for both is that the standard error for the within person experiment is less than that of the between person. This is great. The other interesting thing is that if theta () changes, then this standard error will be the same. Yikes! So smaller effect with the same standard error means a weakened confidence in the effect. You can only go so far with modeling and design and must move into understanding the causal pathway.
In Andrew’s words
Except in the simplest settings, setting up a fake-data simulation requires you to decide on a bunch of parameters. Graphing the fake data is in practice a necessity in order to understand the model you’re simulating and to see where to improve it. For example, if you’re not happy with the above graphs—they don’t look like what your data really could look like—then, fine, change the parameters.
In very simple settings you can simply suppose that the effect size is 0.1 standard deviations and go from there. But once you get to nonlinearity, interactions, repeated measurements, multilevel structures, varying treatment effects, etc., you’ll have to throw away that power calculator and dive right in with the simulations.
Citation
BibTex citation:
@online{dewitt2018
author = {Michael E. DeWitt},
title = {the power of fake data simulations},
date = 2018-09-24,
url = {https://michaeldewittjr.com/articles/2018-09-24-the-power-of-fake-data-simulations},
langid = {en}
}
For attribution, please cite this work as:
Michael E. DeWitt. 2018. "the power of fake data simulations." September 24, 2018. https://michaeldewittjr.com/articles/2018-09-24-the-power-of-fake-data-simulations