Message/Author 


Hello, I have two questions regarding the use of sampling weights with Mplus: (1) If I select a subsample of data, say on a characteristic (e.g., ethnic group), then the sample weights for the subsample will likely not sum to the sample size of the subsample. Is Mplus' rescaling of the weights sufficient or should I recalculate the sample weights for this subsample based on the probability of selection (or is this what Mplus is doing when it rescales the weights). (2) If I wish to look at a multiple group model where the grouping variable is cohorts (actually aggregated over paired, adjacent cohorts, e.g., 1980&1981, 1981&1982, ...). I have a sample weight variable that is computed based on the probability of selection *within* cohort. Because I want to (a) aggregrate adjacent cohorts and (b) then test a multiple group model where these aggregate 2yr cohorts are the grouping variable, can I the weight variable calculated within cohort or will the weights be incorrect for my analysis because of the aggregation and multiple group analysis? I hope that this question is clear Thanks, Scott 


1. You can use the new SUBPOPULATION option of the VARIABLE command in this situation. See page 403 of the new user's guide which is on the website. 2. It sounds like you have a sample for which weights were determined and that this sample consists of cohorts and that you want to place more than one cohort in a group. So the data would be as follows: group cohort 1 1 1 2 2 3 2 4 If this is the case, then I think the weights are fine as is. 


Thank you Linda for your response. I have one followup question. In order to use sample weights, do I need to specify TYPE=COMPLEX? I had originally only specified the weight variable using the WEIGHT IS command. But then I read on p. 205 of the user's manual, "With sampling weights, parameters are estimated by maximizing a weighted loglikelihood function. Standard error computations use a sandwich estimator. This approach can be obtained by specifying TYPE=COMPLEX in the ANALYSIS command in conjunction with the STRATIFICATION, CLUSTER, and/or WEIGHT options of the VARIABLE command." So I then tried to run my analysis with the TYPE=CLUSTER specified, but received this error message: *** ERROR in Variable command TYPE = COMPLEX analysis requires CLUSTER or STRATIFICATION option. Thank you, Scott 


No, weights can be used without TYPE=COMPLEX; COMPLEX requires a CLUSTER or STRATIFICATION variable. 


I have a more 'finicky' problem with sampling weights. I am using NELS data, and the values for some weights are very large (10531.0592 & 11290.0289, for example). [I'm interested in using the SUBPOPULATION command along with cluster and stratification features.] My understanding is that size of these weight values exceeds the column width when the file is saved as a .dat file to be read into MPlus. Inspecting the data file in WordPad reveals that the size of these two weight values causes subsequent values in the row to be pushed to the right (so the value for variable A is read as the value for variable B, for example). When I try to run analyses in MPlus with the full values for these two weight variables, the data are not 'read in' correctly. So, I tried trimming the length of these variables by trimming the weight values of only these two cases to the tenth place (smallest value that would not cause problems in the .dat file). However, by doing so, I receive a message that "SUM OF GIVEN OBSERVATION WEIGHTS IS 10128.06 IT DOES NOT AGREE WITH THE NUMBER OF OBSERVATIONS 10123" There might be a simple solution to this problem, but it has escaped me. Any suggestions? 


I assume that you are saving the data with a program other than Mplus. You should be able to adjust the format of the saved data to accommodate the width of your weight variable. The error message your report is not something I have seen. Please send your input, data, output, and license number to support@statmodel.com. 


I think this problem *might* be because I had wlsmv as my estimator and all dependent variables in the model as continuous. Per p. 342 of MPlus user's guide version 3, this is not an available option. I had been saving the data in .dat format and inspecting the data in WordPad. I'll look into accomodating the width of the weight variable as well. 

J.W. posted on Wednesday, November 18, 2009  12:42 pm



I would like to select a subsample(e.g., males only) from a data set for modeling in Mplus. The SUBPOPULATION option can be used for sample selection with TYPE=COMPLEX. My data set is not from a complex survey. Is there an option that allows the subsample selection. Many thanks for your help! 


USEOBSERVATIONS 

J.W. posted on Wednesday, November 18, 2009  1:58 pm



oK, I have figured out. USEOBSERVATIONS option works. Thanks. 

John Transue posted on Wednesday, October 26, 2011  6:14 pm



My question is very similar to the one that started this thread, but the solution in this thread are not working. I am using ANES data that has a poststratification weight variable, but no clustering. I am using Mplus 5.2. I want to run an analysis on just the white respondents, but I am concerned that if I use the USEOBSERVATIONS the weight will no longer be correct. When I use these commands in the Variables section: Weight is c1_weigh ; Subpopulation IS c1_ppeth == 1 ; Mplus aborts, saying: *** ERROR in VARIABLE command The SUBPOPULATION option can be used only with TYPE=COMPLEX. So then I add TYPE=COMPLEX: Weight is c1_weigh ; Subpopulation IS c1_ppeth == 1 ; Analysis: TYPE=COMPLEX ; When I run this, I get this error: *** ERROR in VARIABLE command TYPE = COMPLEX analysis requires CLUSTER or STRATIFICATION option. Above this post, it says to use USEOBSERVATIONS. That does indeed allow the model to run, but how can I be sure that the weights are used correctly? 


SUBPOPULATION is needed only when you have strata or clusters so with TYPE=COMPLEX. If you have only sampling weights, you should use the default TYPE=GENERAL and USEOBSERVATIONS. 


I have a dataset with about 1000 Ss in 165 schools. At both levels units were selfselected (all schools in the population were contacted). I have information on how many teachers are in each school and thus on the chance that they were selfselected, and also on the chance that a school was selfselected based on population characteristics (region and other). So I created two weight variables, 'weight' and 'bweight'. Is that correct till here? When performing a TYPE = TWOLEVEL analysis I can see that the estimation of the dependent variable (opinion about school reform) is clearly influenced by the 'weight' variable, but not by the 'bweight' variable. Can you help me? 


Sorry, the last sentence in my previous querry should be the other way around: the dependent variable is influenced by bweight, not by weight. 


It all sounds correct. The weight variable probably doesn't have much variation. That is one possible explanation about why you don't see a difference. In general the weight variable should affect the results. 


Thanks already Tihomir! The strange thing is that when I compare this analysis (with weight and bweight variables) with an analysis in which I delete the weight variable (only bweight left), then I got exactly the same numbers (also in last decimals). Do I have to use the wtscale command? When I do that with unscaled, the results change. Maybe I did something wrong in constructing the weight variable. This is how I did it: when 10 teachers participated in a school with 70 teachers, I assigned the weight of 7 for each of the participating teachers. Actually there is a lot of variability in this weight variable. Here's my syntax: VARIABLE: .... CLUSTER = school; WEIGHT = w1; BWEIGHT = w2; ANALYSIS: TYPE = TWOLEVEL; MODEL: %WITHIN% item21; %BETWEEN% item21; 


W1 as constructed says that within school the teachers are equally likely to be selected, i.e., it has no information, and it should not affect the results. If you were running a single level model (not a twolevel model) w1 would be meaningful because it would reflect the probability of selection across the entire population. 


This article may clarify the usage of weights in multilevel models https://www.statmodel.com/download/asparouhovgmms.pdf 


I see. Then what is the appropriate way to deal with the selection bias in this study, considering that both schools (choice by the principal) and teachers who are more in favor of (or against) the school reform may have been more inclined to participate? Thanks again! 


I think your data set might not have information on that. If you have background information for all the teachers in the school (both those that respond and those that don't respond) you may be able to stratify the weights in a meaningful way. Search the literature on selection bias for more ideas. 


Actually I have information on the proportions of sex, age, years of experience in the entire population. Because these variables are also related to the dependent variable, I assume I can use them in creating the w1 weights (post stratification). I have two more questions: is the MPlus approach in dealing with selection bias akin with the Heckman approach in econometrics, and what exactly is meant by "approximately" unbiased estimation method in the paper? 


I would not attempt using weights to solve this problem. I would do two things. First, see how different from the population your sample is with respect to sex, age, etc. Second, I would use those variables as covariates in your model. 

deana desa posted on Wednesday, July 31, 2013  4:23 am



Dear Drs. Muthen and Tihomir, I read this paper, http://statmodel.com/download/Scaling3.pdf, and I have two questions as follows: 1. If an MGCFA model is specified with raw weight (e.g., W1) in conjunction with TYPE IS COMPLEX and CLUSTER IS IDCLUS, does this means Mplus rescaled W1 and thus the weight is add up to the cluster's sample size? Am I misunderstood the default by Mplus when defining weight with COMPLEX and CLUSTER? 2. If the rescaled weight is calculated prior to Mplus analysis (i.e., RW1) and used in the analysis. Is it appropriate or not to define TYPE IS COMPLEX and CLUSTER IS IDCLUS? 3. Is it appropriate to use RW1 without COMPLEX and CLUSTER, that parameters are estimated w.r.t. cluster sample size? 


1. No. This scaling is done for twolevel models not for type=complex. 2 & 3. I would not recommended to rescale the weights with type=complex (if you rescale you can get biased estimates). You should still use type=complex because it accounts for the nonindependence of the observations from the same cluster. 

deana desa posted on Thursday, August 08, 2013  2:02 am



Thanks, Tihomir. My followup question is the following: If TYPE=COMPLEX is used for a single level MGCFA, is there any default scaling that occurs with sampling weights in Mplus 7.11? That is, are raw weights treated as they are? 


Sampling weights are rescaled to sum to the number of observations in the data set if needed. 

Sarah posted on Wednesday, July 16, 2014  8:32 am



Dear Drs. Muthen I have a query regarding weights / complex samples. My model involves the application of a weight and the use of the complex sample feature to account for clustering and stratification. In addition I am looking at a subsample. When I ran my model initially I accidently used the command USEOBSERVATIONS as opposed to SUBPOPULATION. The model ran but I received the warning that the standard errors may be incorrect and to use the SUBPOPULATION command. Upon doing this my model would not run and I received the following message: THE MINIMUM COVARIANCE COVERAGE WAS NOT FULFILLED. USE THE USEOBSERVATIONS OPTION INSTEAD OF THE SUBPOPULATION OPTION TO SEE THE OUTPUT OF THE COVARIANCE COVERAGE. CATEGORICAL VARIABLE GENHLTHR HAS ZERO OBSERVATIONS IN CATEGORY 0. I’m not sure if the zero observations issue is what is causing the problem or not but I’m slightly puzzled as GENHLTHR does not have zero observation in category 0 when I run the model using the ‘useobservations’ command. I’m unsure how to address this problem and wonder if you could kindly point me in the right direction? Many thanks for your help. Sarah 


Please send the data and input file to support@statmodel.com 


Hello. In my analyses, I have students nested in teachers, with probability weights available for each. In general, when we apply student probability weights to studentfocused models, we are able to generalize the results to the population of students represented by the sample. Conversely, when we apply teacher probability weights to teacherfocused models, we are able to generalize results to the population of teachers represented by the sample of teachers. In twolevel modeling where we are able to apply probability weights on both levels (Mplus' WEIGHT and BWEIGHT), what sorts of generalizing statements can be made? Do the results thus generalize to (reflect the population of) both students on level1 and teachers on level2? So either type of generalizing statements may be made in interpreting the results? Thank you. 


Yes. 

Samuli Helle posted on Wednesday, November 25, 2015  2:49 am



Hi. I’d use some advice how to best take the sampling characteristics of my data into account in Mplus. I have data from four parishes. For each parish, I should have all the individuals born since year x (i.e. a census type of data) but before that year I have just samples from each parish. Should I use sampling weights or is there a better way to handle data like this in Mplus? 


Sampling weights looks reasonable. 

Samuli Helle posted on Wednesday, December 09, 2015  4:22 am



Thanks Tihomir! Should the calculation of sampling weights take missing data patterns across strata into account? 


No. Just code the missing values in the usual Mplus style, i.e., code the missing values in the data as 999 and use this option "missing = all(999);" in the variable command. 

Samuli Helle posted on Wednesday, December 09, 2015  10:43 pm



Ok, Thanks a lot! 


Hello, I am wondering which weighting/scaling method I should use in two and threelevel analysis. I have data that were collected through a twostage stratified cluster sampling procedure (students nested in schools). My database provides a final student weight, which is the product of a student weight and a school weight (but I cannot get access to these two components). I know that the school weight was computed as the product of the inverse of the probability of selection of the school and the school nonresponse adjustment. The student weight was calculated as the product of the inverse of the probability of selection of the student from the sampled school and the student level nonresponse adjustment. Now since I have only their products, is it reasonable to use this product as a weight in multilevel analysis? And if so, is unscaled the appropriate scaling method? I would be happy for any recommendation on this. 


The product weight is designed for single level analysis so consider using type=complex instead of type=twolevel analysis if you can not obtain separately the within and the between level weights. If you want to use type=twolevel and you can not obtain the within and between level weights separately your best bet would be to specify the product weight as a within level weight (weight=) and use Mplus default scaling method (so don not specify a scaling= command). This approach takes care of student level weighting but not the school level weighting. Further information can be obtained in http://statmodel.com/download/Scaling3.pdf and http://statmodel.com/download/asparouhovgmms.pdf 


Thanks for your answer. I figured that I could obtain the replicate weights, but I actually want to do multiple group comparisons and examine the role of school level variables (my focus is not on individual level variables). To my understanding, multiple group analysis is not implemented in type=complex, is this correct? 


Multiple group is available with type=complex. 


Thank you, I must have mixed that up then. 

Dan Berry posted on Friday, April 14, 2017  8:36 am



Hello, I’m trying to wrap my head around the following: I’m fitting a twolevel model to longitudinal data. It is specified to estimate the withinperson effect of (groupmeancentered) X and on Y (Level 1) and the betweenperson effect of personmean X on Y (Level 2). To adjust for a large set of potential timeinvariant confounds in the betweenperson estimate, I’ve created a set of inverse propensity score weights for personmean X. I include these weights using the ‘bweight’ function. As expected, the inclusion of the ‘bweight’ impacts the fixed effect and SEs of the betweenperson X effect. I was unclear, however, why it is also having a pronounced effect on the withinperson X effect. I’m clearly missing something about the weighting process. A little help? Thanks in advance. 


You can see the weighting here http://statmodel.com/download/Scaling3.pdf The bweight is designed to take care of sampling weights and to be 1/ prob of selection. So if one cluster has weight 2 v.s. another with weight 1  it is like adding a copy of the first cluster to make it twice as important. Bottom line is  the bweight applies to the entire cluster not just the model on the between level. If you want the within level to be unaffected (which I am not sure that you should ... reading your description it doesn't sound like this is much different from sampling weight) you can try using the opposite weight on the within level 1/bweight in combination with the unscaled option. It could work. 

Dan Berry posted on Saturday, April 15, 2017  11:24 am



Thanks, Tihomir! That clears it up. Best, dan 


Hello, If I would like to apply sampling weight in the multiple linear regression without taking multilevel analysis. Can I do that by adding sampling weight variable and how? Did I get it right that if I apply sampling weight I cannot apply multiple imputation. Best regards, Jintana 


Use Type = Complex (see UG) together with the weight option. 


Hello, I have four stratum, but Mplus allows to define only one stratum? If I have four stratum A=Province, B=District,C=subdistrict, D=village and within village I randomly select respondents. What is your suggestion for defining stratum in the command? Best regards, jintana 


Hello, I run the model with type=complex. The message below appears. I did not face this problem when I apply type=general. Could you may be explain why? THE WEIGHT MATRIX PART OF VARIABLE V IS NONINVERTIBLE. THIS MAY BE DUE TO ONE OR MORE CATEGORIES HAVING TOO FEW OBSERVATIONS. CHECK YOUR DATA AND/OR COLLAPSE THE CATEGORIES FOR THIS VARIABLE. PROBLEM INVOLVING THE REGRESSION OF V ON AI. THE PROBLEM MAY BE CAUSED BY AN EMPTY CELL IN THE BIVARIATE TABLE. THE WEIGHT MATRIX PART OF VARIABLE V IS NONINVERTIBLE. THIS MAY BE DUE TO ONE OR MORE CATEGORIES HAVING TOO FEW OBSERVATIONS. CHECK YOUR DATA AND/OR COLLAPSE THE CATEGORIES FOR THIS VARIABLE. PROBLEM INVOLVING THE REGRESSION OF V ON EDU3. THE PROBLEM MAY BE CAUSED BY AN EMPTY CELL IN THE BIVARIATE TABLE. Best regards, Jintana 


Hello, Please, allow me to explain more. I guess the problems of applying type=complex does not work in this case because some of my items for CFA is rare, in this case item V is the problem. If yes,I would apply type=general instead. Best regards, Jintana 


Answer to your 1:36 question: I think as strata you would consider the total number of subdistricts that you have. And then have: Cluster = Village; assuming that you have at least 30 or so villages. 


Answer to your 2:40 question: Explore the reason for the problem by analyzing only the 2 variables V and AI in the model with V ON ai; Try to restrict your questions to a post in one window. 


Dear Muthen, I would like to ensure whether I get it right. I would type STRATIFICATION=SUBDISTRICT; CLUSTER=VILLAGE; Best regards, Jintana 


Yes, but you have to make sure that your subdistrict variable has unique values for all the districts in all your districts and provinces. Note that this is not really an Mplus question but a complex survey data question  you need to understand that literature to do this correctly. 


Dear Muthen,, Thank you very much for your response. I would appreciate if you could explain me more or suggest further reading. In my case, I have sampling weight, weight variable that is accounted for error due to the complex design. In my dataset, there are five variables 1) province: 5 provinces(five provinces coded 15) 2) district variable (3 districts per provinces coded 115) 3) subdistrict (2 subdistrict per districtcoded 130) 4) village variable ( 3 villages per subdistrict coded 190). I am not sure how to ensure that "subdistrict variable has unique values for all the districts in all your districts and provinces." Normally, if we apply sampling weight in STATA for example, we would apply it without identifying stratum in the command. In this case with Mplus, when we allow to identify stratum with only one variable, I may need your support. Best regards, Jintana 


You were asking about Strata and therefore I answered about the unique values. But now you are talking about sampling weights which are different from strata: "if we apply sampling weight in STATA for example, we would apply it without identifying stratum in the command" I don't know if you understand that these concepts are different  Mplus also does not force you to refer to strata if you are referring to weights. Good complex survey books include Lohr's book Sampling: Design and Analysis. Perhaps you want to talk to a statistical consultant about this. By unique strata I mean that you can't have a subdistrict with the same value in the different districts (for example)  just renumber them. 

SRL posted on Thursday, December 21, 2017  11:57 am



I am trying to run a moderation analyses including sampling weights but am getting error messages. I have a "weight is" statement and my analyses are" type=general; estimator = ML; bootstrap = 10000." Do you have suggestions on where I have errors? 


Send your output to Support along with your license number. 


I am running a sequential mediation path model with TYPE=COMPLEX. I have about 500 level 1 cases (subordinates) working for ~280 unique leaders (cluster). I am trying to calculate the sample statistics and correlations (seemingly the easiest part!) among study variables. However, the univariate statistics for the Leader variables show an N of 500. Is there a way to compute the leaderlevel stats (means, sd's, intercorrelations) and the subordinatelevel stats with the appropriate N's? 


Run Type=Basic twolevel to get these statistics. 


Thank you (as always) for the speedy reply! The complicating factor is I have a lot of missing data and I am letting MPlus handle it (by default). For instance, many subordinates responded (N=1925) but their supervisors did not (and vice versa). So while I have performance ratings for most of these subordinates, they aren't all included in the final analysis. Only 500 or so are (because matched data is available for only ~280 of their leaders). When I run type= basic twolevel, my N's for the sample stats correspond to the overall N, not the final N that I use in my dataset if that makes any sense. Is there a way to only calculate these statistics for the "final sample". Listwise = on doesn't work, because that further reduces my leaders to N~220 and my subordinates to N~400. Thanks so much (and sorry if this is confusing!!) 


O probably misunderstand, but why don't you run Type = twolevel basic on the same "matched data" that you use in your analyses? 

Dan Berry posted on Thursday, March 08, 2018  1:45 pm



Hello. Perhaps a silly question. But I was hoping to get your thoughts on the following re: rescaling weights. I have multiple sampling weights for a given betweenperson unit (there are multiple weights b/c they’re actually inverse propensity score weights, where the selection model changes over time). I need to aggregate these weights, so that there is one aggregate weight per unit (which I’d then include as a “bweight” in a multilevel model). Taking the product of the raw weights gives me some crazyhuge aggregate weights. I’d like to avoid this. My question: would it undermine the utility of the weight, if I first rescaled each of the weights like Mplus does (e.g., rescale_wj = wj (n/sum of wj) and then take their product to create and aggregate weight? Thanks in advance for any thoughts. dan 


As long as you have the same number of weights per person the scaling doesn't matter since the aggregate weight would be scaled by the same constant across persons. If that is the case, the approach you suggest is reasonable. As long as you have the same number of weights per person, I think the way you rescale the weights before multiplying won't affect Mplus results (apart from round off error), so it is best to get the weights as close to 1 as possible (which is what the approach you suggest does). 

Dan Berry posted on Friday, March 09, 2018  6:55 am



Thanks, Tihomir. Very helpful (per usual). 

Jinxin ZHU posted on Monday, April 02, 2018  7:37 pm



May I know why the sampling weights are "rescaled so that they sum to the total number of observations". Are the results affected by the rescaling process? Can I release this default setting, as sometimes the sampling weights are sum to the size of the target population? 


They are rescaled for convenience to avoid values that are too large or too small. The scale does affect the result, with the exception of the loglikelihood which you can adjust that manually by multiplying by (size of target pop / size sample pop). 


There is a typo above  it should say The scale does NOT affect the result ... 

Jinxin ZHU posted on Tuesday, April 03, 2018  6:30 pm



Thank you so much. A followup question: How about the multilevel analysis. I found the results are slightly changed with WTSCALE = Unscaled, when I use the student weight at the within level using TIMSS data. 


It is completely different in multilevel analysis. Take a look at these two articles: http://statmodel.com/download/asparouhovgmms.pdf http://statmodel.com/download/Scaling3.pdf 

Jinxin ZHU posted on Thursday, April 05, 2018  6:46 pm



Thank you so much Tihomir! These articles are really helpful. Currently, I am handling the ICCS 2009 and ICCS 2016 data. Are there any papers/examples of how I should specify the sampling weights in a twolevel analysis using largescale international data sets (such as ICCS, PISA, and TIMSS) with Mplus? Thanks again! 


The above paper is our latest methodological writing on that topic. If you want to see applications you might find some of these useful (back references) https://scholar.google.com/scholar?hl=en&as_sdt=5,45&cites=4068178252377360914&scipsc=&q=&scisbd=1 

Dan Li posted on Friday, May 25, 2018  8:50 am



Hello, I have a question regarding estimates based on weighted and unweighted samples. I used PIRLS data to do Latent Profile Analysis. I specified weight variable as WEIGHT=SCHWGT; In the output: Number of observations 370 It seems that Class Counts and Proportions are based on unweighted sample (n=370) In the LPA, I also included several auxiliary variables. My question is are the estimates (i.e., mean S.E.) for the auxiliary variables across classes based on unweighted sample or weighted sample. Thank you. 


The weights are standardized to sum up to the total sample size. All the analyses use the standardized weights. 


Hello, I am working on an analysis similar to John Transue, above (October 2011 post). I am running a logistic regression analysis on data with sampling weights. There is no clustering. The user's guide says that for looking at stratified data with TYPE=COMPLEX, "Standard error calculations use a sandwich estimator." However, I am wondering what happens when NOT using TYPE=COMPLEX. My question is: how does Mplus adjust the standard errors when using the "WEIGHT IS" command, with TYPE=GENERAL, and ESTIMATOR=MLR? 


It also uses the sandwich estimator  formula (2) http://www.statmodel.com/download/webnotes/mplusnote72.pdf 


Thank you so much Dr. Asparouhov! One clarification  my model does not contain any latent variables, only manifest variables. The sandwich estimator will still be used in this case? 


Is there a way to use different weight variables for different response variables? I mean if I have e.g. response variables DV1DV3 and weight variables W1W3. By looking at the UG, this cannot be done (could multigroup setup in a long format solve the problem...). 


Maybe you can do that, however, if you have different sampling weights due to nonresponse over time we recommend that you use missing values to deal with that instead of adjusting the weights. So using W1 only and specifying missing values for nonresponse is the theoretically correct way of doing this. 


Ok, thanks. If I use multigroup approach does the weight variable influence all the response variables like factor indicators in addition to the main DVs of interest? 


It does but I am not sure how this is related to the previous question. In principle you can use a univariate twolevel analysis where each cluster has size 3 and you can then include the three weight variables but this probably doesn't resolve the underlying issue, which is that sampling weight=1/probability of selection, i.e., inherently the sampling weights should be one. If you are using a multiple group approach I don't think you can have f by dv1dv3 because there will be 3 groups and the variables would be uncorrelated across groups. 


Hello, I have added weights to my latent profile analysis, and the output does not have any errors re: weights, but my profile results are not weighted. I.e. the proportions of each profile is relative to the sample and not the weighted sample. Do I have to specify something other than weights = to be able to get weighted estimates? Thanks! 


How did you make the conclusion that they are not weighted? They should be weighted. You might have to send you example and data to support@statmodel.com if you can't resolve it. It could be useful to add the command savedata: file is 1.dat; save=cprob; You can then see in details how observations are weighted. You can match the weighted posterior probabilities from 1.dat to the results in this section FINAL CLASS COUNTS AND PROPORTIONS FOR THE LATENT CLASSES BASED ON THE ESTIMATED MODEL 


Hello. I have a question about 3level growth model using 6th wave longitudinal study. The 1 level is "within individual(6 repeated)", 2 level is "b/t individual" and 3 level is "b/t school" There are level 2 weight and level 3 weight separately, but no level 1 weight. How can I calculate level 1 weights? Can I set it all to 1/6? And if there is a missing on level 1 due to panel attrition(dropout), do I need to calculate the level1 weight considering the nonresponse? Thank you in advance. 


I would recommend that you start with User's guide example 9.12, i.e., level 1 is given in wide format. Then specify the level 2 weight with the weight command and the level 3 weight with the bweight command. Attrition should be entered as missing data. 

shaun goh posted on Tuesday, May 19, 2020  3:51 am



Hello, I am thinking of measurement invariance testing for a questionnaire across two samples. One sample has weights to correct for selective dropout and design features. The other sample does not have weights. grouping = sample (1 = weighted_sample_A 2 = unweighted_sample_B); weights = var_X; Assuming the above syntax is correct, I was wondering what 'var_x' should contain. Is it simply the weights for each individual from sample A and missing flags for individuals from unweighted sample B? Thank you! Shaun 


var_x should first be standardized so tha the total weight is the same as the sample size, i.e., multiply the original weight variable by (sample size of A)/(sum of all weights). The weight in group B should be set to 1. 

Back to top 