Replies: 2 comments 7 replies
-
I don't have experience with this, but you should probably call |
Beta Was this translation helpful? Give feedback.
-
Following up: so I think my interpretation of thr inferences function for bootstrapping has been wrong. For instance, let’s say I run one imputed dataset but call inferences with 2 draws (R=2) to get the cate….. I would have assumed that for two draws I would get two separate estimates per bootstrap sample per value of each subgroup value I am predicting on…..but they are the same estimates per bootstrap within the same imputation. I assumed each bootstrapped estimate would be different per bootstrapped dataset.. Maybe there’s something I’m missing. Any advice would be great. Thank you again! |
Beta Was this translation helpful? Give feedback.
-
Dear group,
I am working on estimating the conditional average treatment effect (cATE) for a set of analyses. In other words, I am estimating the exposure effect for each level of a variable of interest in my model using G-computation.
One of my sub-pieces is going beyond complete case and integrating multiple imputation. Basically, the challenge I am having is I am doing bootstrapping and MI together under the context of estimating the cATE. I have tried this two broad ways as per the literature:
Approach 1: I have tried everything and will share my code but it has been a struggle:
#Estimation of model across each iteration of the multiple imputation dataset (about 500 iterations for now)
#Going across each imputed dataset and doing avg_comparisons to generate my cATE while bootstrapping and storing bootstrapped results for the cATE for each MI iteration.
storeit=list()
#just trying a little test (also used lapply, etc but just being transparent)
for(i in 1: 5){
check=avg_comparisons(lm(out.formula,data = dat[[i]]), variables="exposure", by=c("variable of interest to estimate exposure effect across"),newdata = datagrid("Vector of values for by variable above" = ValVec)) %>%inference(method="boot")
storeit=rbind(storeit, check)
}
**My first question:
Approach 2: I created a bootstrapped dataset from my full dataset (which included missings) and then impute twice per bootstrap iteration before trying. In the middle of this actually but may have follow-up code later this week on it to compare as I think the confidence interval coverage may be a little better ultimately.
Any experience on either of these again would be appreciated so I don't have to do everything in nuts and bolts (would love to do in context of marginaleffects).
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions