Message/Author 

Stan Hong posted on Wednesday, December 14, 2005  12:49 am



Dear Linda/Bengt Using Mplus, is it possible to test mediation effect in discrete time survival analysis? I'd like to test X>M M>f X>f, where X=independent variables and M=mediators. Are the point estimates and standard errors of the indirect effect generated by Mplu3 valid in this case? Thanks so much in advance. 


Yes and yes. 

Stan posted on Saturday, December 24, 2005  8:13 pm



Merry Christmas! Can you please let me know the formula of its standard error (for mediation effect in categorical data analysis)? Which formula is used for computing a standard error of mediation effect in Mplus in general? 


For indirect effects in Mplus, standard errors are estimated using either the Delta method of bootstrap. 


Can you refer me to examples of testing mediation effect in discrete time survival analysis? Thanks! Lois 


I don't know the answer to this question. I will try to find the answer and post it if I do. 


Jenn Tein @ ASU has a chapter that should be relevant: Tein, J.Y., & MacKinnon, D. P. (2003). Estimating Mediated Effects with Survival Data. H. Yanai, A. O. Rikkyo, K. Shigemasu, Y. Kano, J. J. Meulman (Eds.) New Developments on Psychometrics (pp. 405412). Tokyo, Japan: SpringerVerlag Tokyo Inc. I'd contact her or Dave MacKinnon directly if you have trouble tracking down the book....... 

Steffi posted on Thursday, June 22, 2006  2:29 pm



I am afraid my problem is very basic but I cannot come up with a good explanation. Running a DTSA with four binary indicators (x1x4), and two continuous predictors (y1 and y2) the model quickly converges (Mixture Missing) and yields reasonable results. The relevant Model statement is: %overall% eta BY x1@1 x2@1 x3@1 x4@1; eta ON y1 y2; [eta@0]; The predictors y1 and y2 are correlated, but adding (y1 WIH y2) requires numerical integration, which never converges. The same is true for a mediator effect (i.e., eta ON y1; y1 ON y2;). The data are nice, and y1 and y2 are clearly correlated (and a mediator effect seems likely), so I think it is something more fundamental I am missing here... Thank you very much for any insights! 


The predictors y1 and y2 are correlated by default in the Mplus analysis. Just like in regression analysis, these correlations are not part of the model's parameter set. You should not mention y1 with y2. If you really want the mediation model that you have problems with, contact support. 

Steffi posted on Thursday, June 22, 2006  11:26 pm



thank you for your fast reply! Indeed, the mediation is what I am actually after. But out of curiosity: other than in "standard" CFA I assume it is not possible to constrain corr(y1, y2) to zero? THANK YOU 


I think you are asking if in a mixture analysis you can constrain the correlation of two variables to zero if they are predictors of a continuous latent variable that in turn influences categorical observed variables. That is possible in Mplus. If you have problems doing so, please send your input, output, data, and license number to support@statmodel.com. 

Steffi posted on Monday, July 24, 2006  7:51 am



I have a follow up question on Stan Hong, Lois Gelfand and my earlier post: I want to test a mediation effect in discrete time survival analysis (X>M>Y) with X = continuous predictor, M = continuous mediator and Y = vector of binary event indicators (proportional hazard odds assumption). I have censored observations, and from my reading of the literature, the best approach would be the product of coefficients method (a*b, where a is a standard regression coefficient and b is a logit coefficient) using bootstrapped SEs. With a logit link function and ML, am I correct that such a test is not (yet) possible in Mplus? The next best approach is probably to compute SE using the Delta Method as recommended by Linda somewhere, but this also needs to be done by hand, correct? Before tackling this task, I want to make sure that there is no "better" alternative (such as a probitlink function with WLSMV as indicated in another post, etc. but I am very hesitant to adopt such an approach since I have never seen it before in a DTSA context). What would you recommend me and can you provide me with any good references for mediation effects with dichotomous outcomes (I am aware of MacKinnons message from Dec. 04 and many of his excellent papers, but I wonder to which extent these findings can be generalized to DTSA with censored observations)? Thank you so much in advance! 


You are right that we don't do model indirect for categorical ML. But you can use model constraint: new(c); c=a*b;  this gives the Delta method. WLSMV is not appropriate with survival indicators given that missing data needs to be done via MAR. I can't think of any mediation references with survival, but you may want to contact Katherine Masyn about that. 


I want to do discretetime hazard model. My data set is in long formatobservations at different ages are nested within individuals, which are nested in cluster (census tracts). Can I specify id of individuals as "idvariable" and census tracts as "cluster" to do three level model? Or do I have to restructure the data so that each age interval becomes one variable and to take a multivariate approach at the lower level? Also, can I introduce frailty to account for unobserved heterogeneity for individuals without restructuring the data? Any comments will be greatly appreciated! 


This can only be done in a wide format. Frailties can be introduced in the wide format. 


I am trying to use the "Model Constraint: New" command described by Bengt above on July 26, 2006 4:25 pm to obtain the Delta method for estimating the standard error of a mediation effect in a cox regression model. However, I get a fatal error message: "*** FATAL ERROR Internal Error Code: PR1004  Parameter restriction split problem. An internal error has occurred. Please contact us about the error, providing both the input and data files if possible." Am I doing something very wrong here? My syntax is below: Usevar= RGRLEV I_123Y P_123Y FAMPROB6 ; survival=P_123Y (all); timecensored=I_123Y (0= right 1=not); Model: FAMPROB6 on RGRLEV (a); P_123Y on FAMPROB6 (b); P_123Y on RGRLEV ; Model Constraint: New (c); c=a*b; 


The problem is that one variable is continuous and the other time censored. This is not allowed in MODEL CONSTRAINT. If you are creating an indirect effect, I am not sure that the product can be used here. You can ask for TECH3 and compute the standard error of the product yourself. 


HI Linda, Thanks for your reply on Nov 8. Could I please confirm with you about not being sure if the product can be used in the example syntax I posted on Nov 8? I would ideally like to estimate the indirect effect of RGRLEV on P_123Y (survival time). However, this would mean multiplying the estimates of the regression of FAMPROB6 on RGRLEV, and the cox regression of P_123Y on RGRLEV. As these are different kinds of regression analyses, perhaps I cannot multiply the paths together to obtain an indirect effect is this correct? Is there any way around this? I suppose I could forego the cox regression model and estimate the logistic model regressing I_123Y (survival or not) on FAMPROB6  and then multiply the path from RGRLEV to FAMPROB6 and then FAMPROB6 to I_123Y to obtain the indirect effect of RGRLEV on I_123Y. 


See the following paper: MacKinnon, D.P., Lockwood, C.M., Brown, C.H., Wang, W., & Hoffman, J.M. (2007). The intermediate endpoint effect in logistic and probit regression. Clinical Trials, 4, 499513. 


I have two questions concerning discretetime mediated survival analysis. Our data set is in long format (observations nested within individuals). The predictor is dichotomous, mediators are continuous, dependent variable is an event indicator (no missing data). 1. Our model is a survival analysis with a preceding panel regression model. Is it correct to use TYPE=COMPLEX in conjunction with the CLUSTER option to account for the nonindependence of observations (due to the long format)? Or are there any alternatives you would suggest in our case (any literature suggestions are appreciated)? 2. Is it true that indirect effects can only be estimated with a probit regression in our case? Is there any alternative model specification based on the standard logistic link function? Thank you in advance, Oliver 


1. It sounds like you have longitudinal data ("panel regression model" followed by a subsequent survival model. That seems to be best handled by letting the panel part be in wide, not long form (singlelevel analysis) since the survival modeling is in wide form (so first the columns with the panel outcomes for the panel time points, followed by columns for the even indicators). 2. You can consider the indirect effect with f as end point as in regular linear regression since f (using the Mplus UG notation) is continuous. 


Thank you so much for your fast response. We can see that your suggestion is a viable alternative. But: We are mainly interested in the effects of timevarying mediators on the outcome. If we use a wide data setup, we get several effects (as many as there are panel waves, in our case 8) on the respective event indicators. We feel this is quite cumbersome to depict/interpret. Would it be correct to impose an equality contraint on the various effects per (timevarying) covariate? In this case there would be only one effect per timevarying covariate. (Still, if the model fits worsens after the constraint, we continue to have the "cumbersome" multiple effects, don't we?) 


You can add equality constraints. Do you really believe that the relationship is the same at all time points? 


Thanks for your answer. We are pretty sure the relationship is not invariant across time. But if this is the case, the results become quite technical/complicated for a nonmethodological journal article, which we are trying to avoid... Besides, isn't the equality constraint just what the "usual" timediscrete event history analysis with timevarying covariates does (unless you include interactions with times of measurement)? One last question: Would you consider the "long format" approach (every row in the data refers to an episode of observation) incorrect, even if nonindependence of observations is controlled for by adjusting standard errors (by using the Mplus CLUSTER command)? In the "classic" (i.e., nonSEM) literature, timediscrete event history analysis is usually done in long format (this is why we figured we could run our analysis that way). 


On the 1st paragraph  you can certainly do it with equality, hoping that fits well. But if it doesn't I am sure you can convince editors of the need for a better albeit slightly more complex model. You can use the long format approach if (1) the nesting is the same for both the growth and the survival part, and (2) you take the older, long approach to survival. 


Dear Linda/Bengt I'd like to test a discrete time survival model with several independent variables (X) and a mediation effect(M). X>M M>f X>f, My syntax is MODEL: f by t5t18@1; f on X1 X2 M; f@0; M on X1 X2; I know that it is possible to test mediation effects with MPlus but my model doesn't work this way. Is there a mistake in my syntax? Do I have to set the variance of f free? Thank you so much in advance! 


This look correct. Please send your output and license number to support@statmodel.com. 


Dear Linda and Bengt, I am doing a discrete time survival analysis by following models described in the Muthén and Maysen's article, and the Mplus user’s guide. However, I am experiencing some difficulties. 1) I have 10 time points but for the first two time points the binary U’s indicators have only 0 values. As the current version of Mplus doesn’t allow a categorical indicator with only one value, I thought of doing the analysis without the first two time points. That strategy can however result in bias in the estimation of the hazard because of the suppression of information related to time interval. Is that reasoning correct? If yes, is there any way to evaluate that bias? If the bias is negligible I can carry on with the discrete time survival analysis especially because the continuous one doesn’t really suit my problematical. 2) In the Muthén and Maysen's article they mentioned imposing specific structure to the logit baseline hazard. So, I was wandering how to test linear trend on the logit baseline hazard via MODEL CONSTRAINT. I am looking forward to hearing your answer. Thank you for your invaluable help. 


1) I may be wrong, but I don't see how deleting the first time points with u=0 biases the results. You simply have survival probability 1 during that time. 2) You can simply use the growth language i s  u1@0 u2@1... and make sure you don't get variances for the growth factors i and s. The use of "f BY u1up@1 is the same as what the i growth factor does. 


I am planning to do a discrete time survival analysis and had a few questions I was hoping you could answer: 1) How should the data be coded for people who have missing assessments but then come back into the study? For example, say I have six assessments and someone drops out at time 3, but then comes back into the study at time 5 and reports the event. Should it be coded: 0 0 999 999 1 999 (999=missing) if this hypothetical person did not report the event at time 5 or 6 would it be coded? 0 0 999 999 0 0 2) How does one handle unequal spacing between time points in these models? Thanks in advance. 


1. In the Muthen/Masyn implementation of discretetime survival analysis in Mplus, only nonrepeatable events such as onset of drug use are considered. Because of that, I think the coding for your two cases would be 0 0 0 0 1 999 and 0 0 0 0 0 0 Missing has a different meaning in this coding scheme. 2. The spacing does not matter unless you put a growth model on the thresholds. 


Linda, Thanks for the reply. Since the people have missing information for some assessments it seems problematic to assume that the event has not occurred in for these phases (i.e., 0). I thought there was a way to take periodic missingness into account in these models. If I understand correctly dropout prior to an event occurring would be handled by coding 999 at the timepoint where dropout occurred, but handling periodic missingness as I outlined previously seems more tricky. 


The question would need to ask if it is the first time an event occurred. I don't think your coding scheme would work in the Mplus implementation. 


I have a dataset where the units under observation are subsidiaries of multinational companies. In a period of 20 years new subsidiaries are setup that enter into the dataset. How should I code for the initial periods when they were absent because they were not born at all? They are not really missing values. Thanks for your help 


Since you posted under discretetime survival, I assume you observe some event such as failure. Perhaps you should consider as your survival time variable, the number of years between being born and the event? 


Yes, we do observe a failure and that is the event of interest, not the founding/entry. As of now I am not considering timevariant covariates (country level contextual factors) but the idea was to incorporate them at a later point of time. If I didn't have these I could have potentially coded each subject from its first year of entry. Even if I did this it would still bias any unobserved heterogeneity for the particular calendar year, for ex: there could have been a bad economic condition during a particular year in a particular country. So I just reasoned that I will code it as missing since it is not part of the risk set in any case. If I correctly understood your suggestion  I can incorporate age as a timevariant covariate. I am just wondering if there would be too much of a correlation between this and the dependent variables (u) and wouldn't that overshadow the effects of my other variables? 


We think age should not be the covariate  it should be the dependent survival variable. Instead we would recommend that the economic conditions that you are talking about should be a timevarying covariate  that can be done as in the http://statmodel.com/download/lilyFinalReportV6.pdf Discrete time survival may be a good approach here that can also accommodate timevarying covariates easily. 


Yes, discrete time survival is the most appropriate approach for more than one reason for my analysis. If I have age as the survival time dependent I cannot have macroeconomic conditions as a simple timevarying covariate. Let me illustrate with an example. In a particular country consider subsidiary 1 as existing between 2006 and 2010. Subsidiary 2 exists from 2004 to 2008. I can have 7 dummy binary variables for age 1 to 7 but neither of them reached age 6. I now have 7 values for macroeconomic condition. The macroeconomic condition of year 2008 was fatal for one and not the other but with a simple timevarying covariate it appears that the fifth year (2008) was fatal for both. If I correctly understood why you provided the particular link, what I can do is incorporate the existence of the subsidiary in a given year and year based timevarying variables in the latent part. This will incorporate the calendar year based effect on the survivability, through the class membership. Am I correct in interpreting what you said? 


The time varying covariate would have to be coded for each observation separately x1, x2, ..., x7 x1 would be bad economic condition in first year of existence x2 would be bad economic condition in second year of existence ... thus the data for the two companies would look like that age, x1, x2, ..., x7 5,0,0,1,0,0,0,0 5,0,0,0,0,1,0,0 Now if you also add discrete time survival  the value 5 will also be split into binaries age1, age2,...,age7 age1, age2,...,age7, x1, x2, ..., x7 0,0,0,0,0,999,999 ,0,0,1,0,0,0,0 0,0,0,0,1,999,999 ,0,0,1,0,0,0,0 


Dear all, we are doing a discretetime survival analysis in which we examine whether the effect of one timeinvariant predictor (proportionality assumed and confirmed) on event occurences is mediated by a timevarying predictor (proportionality NOT confirmed). The basic model looks like this: ! c'path event_T1 on t_iv (1); .. event_Tn on t_iv (1); ! bpaths event_T1 on t_v1 (b1); .. event_Tn on t_vn (bn); ! apaths t_v1 on t_iv (a1); .. t_vn on t_iv (an); To test mediation, we use the model constraint approach and everything works fine. However, given that the timevarying mediator is nonproportional we get n mediation effects (in our study n = 8); this makes sense but to facilitate the interpretation and reporting of the results, I was wondering whether there is a procedure or an approach to average the n effects? Or would you recommend to compute the model with proportionality assumed for the timevarying predictor? Thanks in advance, Steffen 


You can express the average in MODEL CONSTRAINT. 


Hi Linda, thanks for your advice. Is the standard error for the average valid? Thanks again, Steffen 


The SE that comes from Model Constraint is valid. 

Stefano posted on Friday, January 04, 2013  6:33 am



Dear Linda/Bengt, I am running a discrete time survival analysis with a timeinvariant predictor and considering a mediator (proportionality NOT confirmed). Based on the last note on this thread, I would like to obtain the average of the n indirect effects, but I didn´t manage to write the correct code. (simplified example with only 3 times) f BY u1u3@1; !c' path f on t_iv; !b paths u1 on m (b1); u2 on m (b2); u3 on m (b3); ! a path m on t_iv (a); f; model constraint: new (ind1); ind1 = b1*a new (ind2); ind2 = b2*a new (ind3); ind3 = b3*a new (indirect); indirect = mean(ind1 ind2 ind3); How should i specify the model constraint to get the average correctly with their respective standard errors? Thanks for your help! Stefano 


The MEAN option works in DEFINE not MODEL CONSTRAINT. Try indirect = (ind1 + ind2 + ind3)/3; 

Stefano posted on Friday, January 04, 2013  8:07 am



Thanks a lot for your advice; it works this way. How are the standard errors calculated for this indirect effect? Are they unbiased? Thanks again for your help! 


They are Delta method standard errors. They are unbiased in large samples. 

Stefano posted on Saturday, January 05, 2013  8:58 am



Thanks for your reply. Another question: is there a way to get other model fit statistics besides AIC and BIC in a multilevel discrete time survival analysis with mediators? 


No, you would have to compare competing models. 

Cheng posted on Wednesday, September 20, 2017  11:26 pm



May I know can I run Survival analysis for three outcome events (e.g., patient's survival, catheter free survival, and peritonitis free survival) in Mplus 7.4? If yes, time for different events can it be done by using Cox within SEMSurvival analysis? Thank you very much 


See section 3.1 in http://statmodel.com/download/SurvivalJSM3.pdf and the Frailty directory of simulation examples in http://statmodel.com/download/survival16.zip To use Cox all you need to do is specify the variables like this survival=t1 t2 t3; 


I am interested in modeling student permanence in higher education, over a 4 years period using a discretetime survival model. In that sense the event is repeatable every year, which it seems to be not plausible to model in Mplus. However, if I change the event to be the opposite, i.e. to dropout from higher education, it can be modeled. If I estimate the discretetime survival model of higher education dropout. I wonder if I can still interpret these results in relation to student permanence in higher education. Assuming that the effects on permanence would be the opposite of the findings for the dropout model. Or, is there a way around to estimate a repeatable event model in Mplus? I would appreciate any suggestion to model student permanence. 


Have a look at the paper on our website under Papers, Survival Analysis: Masyn, K. E. (2009). Discretetime survival factor mixture analysis for lowfrequency recurrent event histories. Research in Human Development, 6, 165194. download paper contact first author show abstract 


Hi, Alejandro. If you estimate the hazard of dropout, you can get back to survival estimates of dropout (i.e., the probability of "permanence" at time t). Consider that in classical survival analysis, modeling the hazard rates of death allow you obtain estimates of, say, the median survival time. In my understanding of your research question, modeling the hazard of dropout will allow you to estimate permanence (survival) rates across time. With only a four year time period, you could also consider using a simple Markov chain model (in Mplus this would be specified as a latent transition analysis with a single, perfect, indicator latent class model for each year). With this specification, you could extend to a constrained mixed Markov chain model (e.g., "moverstayer" model). Hope that helps. Katherine Masyn 


Thank you Katherine and Bengt for your kind responses, they are very helpful. Katherine, my research is focused on investigating successful higher education permanence for firstgeneration students in higher education. Therefore, my interest on explaining permanence rather than dropout for this group of students. I wonder if I can ask you for a couple of references to your suggestions. I am not sure how to interpret the estimated coefficients of the dropout model in relation to the survival (permanence) of the students. However, I understand that survival rates can be derived from the hazard rates. Can you point to a reference of Markov models for a similar case to student permanence? Many thanks! 


Dear Bengt, I am helping a postdoc to fit a discretetime survival model with mediators. We are using WLSMV with the THETA parameterization in order to obtain global fit stats and also refitted it with BAYES as a sensitivity analysis. The form of the model is: Exposures + covariates > mediators set 1 Mediators set 1 > mediators set 2 Exposures + covariates + mediators set 1 + mediators set 2 > outcome We also correlate residuals of mediators at each mediation stage and bring xvariables with missing data into the analysis. Coauthors are struggling with interpreting the probit coefficients, so I was wondering: 1. Would it be appropriate in your opinion to multiply the probit coefficients by 1.7 and then exponentiate them to obtain an approximate hazard ratio interpretation? (I realize this would be only approximate). 2. Alternatively, could we use the instructions on pp. 552553 of the Mplus UG to compute predicted probabilities at various levels of an exposure and then compute their differences using MODEL CONSTRAINT? If so, would it be sufficient to use predictors of the outcome in a single equation as in the UG example or is the expression for the probability more complex due to the presence of mediators, correlated residuals, and/or bringing xvariables with missing data into the model? Many thanks in advance and best regards, Tor Neilands 


WLSMV should not be used for discretetime survival because it doesn't handle the missing data correctly. Use ML or Bayes. As opposed to WLSMV and Bayes, ML offers logit but may require many dimensions of integration. 


Thank you for this feedback. So noted. We used a personperiod data structure as described in Paul Allison's SAS logistic regression book. We then used WLSMV in order to obtain robust results because some variables in the analysis are continuous, but nonnormal. We could use MLR, but would then need to compute the indirect effects by hand, which we were hoping to avoid. That does have the appeal of using the logit link, as you pointed out, and the event is somewhat rare, so perhaps that would be best statistically. Bootstrap with numerical integration with many points of integration could be quite timeconsuming. What about WLSMV + multiple imputation? Would that be OK? In any event, with WLSMV + MI or Bayes, would it be possible to compute approximate hazard ratios using exp(1.7*probit coefficient)? For MLR or Bayes (or WLSMV+MI if that combination is OK), would it be possible to compute predicted probabilities as described in (2) of my original post? Thanks again and best wishes, Tor 


I assume the discretetime part is the "outcome" that you mention; perhaps it is the "f" of UG ex 6.19. And you might be interested in indirect and direct effects of the exposure on f. Given MLR's heavy numerical integration demands in your case, I would use Bayes. Perhaps a 1.7 (or 1.8) multiplication is reasonable  I haven't tried that for hazards. Regarding your idea in (2), I assume you are thinking about the probabilities of "u1u4" in ex 6.19 terms. Those are influenced by the distribution of f, so you need to get the mean and variance of f as a function of the mediators at different exposure values (which is without complications). 


Dear Bengt, Thank you so much for this discussion and your guidance. Mplus Discussion is such a wonderful resource for enhancing the rigor of our research projects. From our discussion here, it sounds like we should go with our Bayesbased results. I appreciate your opinion on the idea of converting the probit coefficients from the Bayesian analysis to hazard ratios. For the idea of generating predicted probabilities, we're not actually using the UG Ex. 6.19 approach (though we tried it to verify we got very similar results to our approach and we did). We are using the discrete time survival approach described by Paul Allison in his SAS logistic book on pp. 235236, "Each individual's survival history is broken down into a set of discrete time units that are treated as distinct observations. After pooling these observations, the next step is to estimate a binary regression model predicting whether an event did or did not occur in each time unit." So, the outcome (in our case the outcome is contracting HSV2) is treated as an observed 0/1 variable with one record for each discrete time period the person was observed (i.e., a personperiod data structure). I'm not sure whether that would change any of the recommendations and conclusions you offered above? Thanks again for everything, Tor 


I see, so you want P(u=1) for each of your u's (using notation of ex 6.19). That can be expressed in Model Constraint using the PHI function for the standard normal distribution function of probit. But note that the residual variance of the underlying y* conditional on the exposures and covariates is not 1 as in regular probit, but you have to add to 1 the residual variance expressions due to the 2 continuous mediators. A similar exercise is carried out in section 8.1.1 of our RMA book for the case of a single cont's mediator and a binary outcome. 


Hi Bengt, As a followup to this helpful thread, we wound up using ESTIMATOR = BAYES and mediation assessed via the traditional mediation approach with MODEL INDIRECT. We used the approach of multiplying the probit coefficients by 1.7 and exponentiating them to obtain approximate ORs (HRs). A statistical coauthor on the resulting manuscript has raised a question regarding interpretation of the model estimates. He asks: “How can one interpret the regression coefficients in a model with exposure X, continuous mediator M and binary outcome Y? Specifically, can the coefficients in the regression of X on M be interpreted as standard linear regression coefficients (change in M per one unit change in X)?” To those questions, I would also ask how the indirect effect is interpreted  as a linear regression of X on Y* through M or as the probit regression of X on Y through M? We are trying to figure out which coefficients to multiply by 1.7 and expotentiate to get approximate odds ratios and which to leave untransformed as linear regression betas. I hope this all makes sense. Thanks very much in advance for your help and best wishes, Tor Neilands 


Yes, X to M is a standard linear regression. I would think you want the influence of X on Y through M as expressed by the probit regression, that is, taking into account the M residual variance in addition to the probit residual variance of 1  just like in Section 8.1.1. in our book. 


Thanks, Bengt. I am still struggling to understand this. Thanks for bearing with me. So, for the X>M>Y relationship, is the distinction that the product of X>M and M>Y coefficients in the classic mediation approach is really the product of X>M and M>Y*, where Y* is the continuous underlying latent representation of Y? And thus if one wants the effect on Y rather than the effect on Y*, it is not as simple as computing an approximate OR as exp(1.7*indirect effect coefficient)? I have been using X>M>Y as a simplified example. Our mediation model has multiple covariates and mediators, with some mediators being binary and at least one other being continuous. The variables are in an X > M1 mediators > M2 mediators > Y configuration. What would you advise in this type of situation to maximize interpretability of results? Thanks again, Tor 


Q1: Yes Q2: Right Q3: Are you still considering Y as a series of binary variable representing a discretetime DV? 


Hi Bengt, For Q3, conceptually the answer is yes, though in our application Y is a single personperiod 0/1 variable indicating HSV2 incident infection. The data structure follows the structure recommended in Paul Allison's SAS survival regression book in the chapter on discrete time survival. I.e., "long" format with repeated observations for each period per person. Best wishes, Tor 


It's hard to say what the best approach is here  it is a bit of a research topic for several reasons and I don't feel I can answer well without doing research on it. First, you have a binary Y and actually several binary Y's so mediation effects need special counterfactual formulas. Second, you have sequential mediation which needs special counterfactual formulas or be done using marginal structural modeling (as referred to in the "causality" literature). And when one takes the simpler WLSMV and Bayes approaches of using a probit link and considering Y* instead of Y to keep the regressions linear, the issue of how to get back odds ratio effects via 1.7 scaling is untried as well especially for discretetime survival purposes. I wish I could be more helpful. Perhaps SEMNET has suggestions  Cam McIntosh is now actively discussing causality matters. 


Thanks so much, Bengt, on behalf of my coauthors and me. After some discussion, we have decided that we will report the Mplus output directly as linear regressions among the continuous Y's and the categorical Y*'s and explain in the data analysis description in the methods that the results should be interpreted as linear regressions for the Y's and probit regressions for the Y*'s. Does that sound correct? If it is, do you have a favorite way to describe the regression of a Y* variable onto an X? E.g., "For every one unit change in X, the latent variable representing Y changes by <coefficient>"? Also, if you have favorite citations for references that discuss interpreting probit coefficients for Y* variables, those would be much appreciated. Thanks again for everything, Tor 


You consider a linear regression for both M as DV and for Y* as DV (we'll get to Y later). And, yes: "For every one unit change in X, the latent variable representing Y changes by <coefficient>" And you want to say how much that Y* change is in Y* SDs (so use STDYX). But I would also relate that Y* change to the corresponding Prob(Y=1) change. The probability change due to 1 X unit of change is different dependent on where on the Y* scale you are  I discuss this in Section 5.2.6 of the RMA book. 


Thank you, Bengt. This is very interesting and I'm learning a lot. I truly appreciate you making the time to walk me through these steps and hope that our discussion benefits others who might find themselves in a similar situation. I had one followup question to your helpful post above: To relate the Y* change to the corresponding Prob(Y=1) change, I'm wondering how to do that. Would it be permissible to follow the approach shown in the RMA book on p. 232233 (generating the predicted probabilities and plotting them)? I ask because our situation diverges from the example shown in the RMA book in two ways: 1) We're using BAYES with estimation of Y* rather than ML with estimation of Y as in the RMA coal miner example on p. 233 and 2) We're working with a complex path mediation model rather than a regular regression model. (Our previous discussions above about the challenges of computing predicting probabilities for Y given the presence of mediators prompted this question.) It would be fantastic if we could present predicted probability plots as I imagine at least some of our more epioriented coauthors and readers would appreciate a linking of the Y* results to the Prob(Y=1). Best wishes, Tor 


Sorry  one more clarifying question, if I may: For the recommendation to standardize, I can see how STDYX would be the best choice for quantifying associations among two continuous variables (including continuous underlying latent representations of M* and Y* and any continuous X's). If, however, X is binary, then I assume I'd want to substitute STDY for that association, correct? Thanks, Tor 


Here is some more detail on the Y* mean and variance and the connection with P(Y=1): Look at section 8.1.1 in our RMA book, formulas (8.1)  (8.6). This is with probit so relevant for Bayes. This is for one mediator but it can be generalized to several ones like you have. Note the reduced form expression (8.3) and the fact that the V(Y*  X) variance is not 1 as in simple probit regression (where c=1), but also has a term due to the mediator residual variance (you will have more than one such term due to more than one mediator). Do the same reducedform exercise for your case. Note the probability expression in (8.6) which can be computed using the PHI function in Model Constraint. With a binary X and one binary Y, you can even formulate an odds ratio based on the 2 Y probabilities and the 2 X values. This OR won't be constant with respect to control (covariate) values as in logistic regression but the ORs can nevertheless be computed  and then plotted in Model Constraint's PLOTLOOP statements. See also Section 5.2.10 in the book about the fact that probit can also give estimated ORs. Hope this helps. 


Dear Linda/Bengt, I am fitting a discrete survival model across multiple groups with mediation. The aim is to test whether the explanatory role of mediators are different across two groups. I have used examples from the literature and have resorted to an input where I use the knownclass option in order to define groups and model constraint option to define indirect effects %OVERALL% t by TM1TM15 @1; t@0; mas on edu; t on edu ; t on mas ; %bi#1% mas on edu(a1); t on edu (c1); t on mas (b1); %bi#2% mas on edu(a2); t on edu (c2); t on mas (b2); MODEL CONSTRAINT: new(ab1 ab2 diff1); ab1 = a1 * b1; ab2 = a2 * b2; diff1= ab1  ab1; Overall this seems to work. However from one of the answers above I gathered that it is not possible to obtain CI’s via bootstrapping if you use estimator = mlr, is this correct? Are there any alternatives to still obtain bootstrapped indirect effects? What are the downsides of not bootstrapping here and do you know of any literature describing this? Thank you in advance! 


Estimator = MLR means that while estimates are ML, the SEs are computed using a special "sandwich" formula. So it would be illogical to do bootstrapping with MLR because bootstrapping gives its special SEs. So simply say Estimator = ML when you do bootstrapping. 


Thank you very much. I am using ML now. However I still have a follow up question. When I try bootstrapping with ML it gives me the following error: *** ERROR in ANALYSIS command BOOTSTRAP is not allowed with ALGORITHM=INTEGRATION. In my analysis command I use the following: ITERATIONS =10000; estimator = ml; ! BOOTSTRAP = 1000; TYPE=MIXTURE; INTEGRATION = MONTECARLO; ALGORITHM=INTEGRATION; The INTEGRATION = MONTECARLO command was requested by mplus after adding “mediator on x;” or “mas on edu;” in above in the model statement. TYPE=MIXTURE; and ALGORITHM=INTEGRATION; were requested after including the knownclass option for mulptiple groups. With these statements everything works (except for bootstrapping) and results comply with what I found doing the same analysis in STATA or SPSS. However I am still wondering, do these abovementioned statements seem correct or am I doing something wrong? Is there a possibility to still do bootstrapping or is this not necessary when using ml? 


Bootstrap is allowed with Integration in Version 8.4. 

Back to top 