Home › forums › Mixed Models › When to uncorrelate random effects
 This topic has 1 reply, 2 voices, and was last updated 6 years, 7 months ago by henrik.

AuthorPosts


February 16, 2018 at 10:08 GMT+0000 #204João SantiagoParticipant
After reading a substantial amount of literature about LMMs, I still don’t have a good intuition for when to uncorrelate random effects (for example using the double pipe , (a + b  z)
What is a practical example of this? Could be a research question that explicitly calls for this, for example.

February 16, 2018 at 13:42 GMT+0000 #205henrikKeymaster
There probably are research question that explicitly call for suppressed correlation estimate. However, they are mostly suppressed to avoid convergence problems. We wrote in our chapter on mixed models:
For the limited sample sizes that are common in psychology and related disciplines a common problem is that the maximal model is not fully identified (Bates, Kliegl, Vasishth, & Baayen, 2015), especially for mixed models with complicated random effects structures. Even though the optimization algorithm converges to the optimum (i.e., the maximumlikelihood estimate for a given data set) the variancecovariance matrix of the random effects parameters at the optimum is degenerate or singular. At least for models estimated with lme4 this is often signified by convergence warnings. Other signs of singular fits are variance estimates of or near zero and correlation estimates of $\pm 1$. The occurrence of such situations is due to the fact the parameters associated to random effects (e.g., $\sigma^2_{S_\delta}$) are more difficult to estimate than fixed effects (e.g., $\beta_{\delta}$).
The important part for your question is the last sentence. Randomeffects parameters are more difficult to estimate than fixedeffects. We need considerably more data to estimate a variance or a correlation parameter than a simple fixedeffect. Furthermore, correlations require even more data than variances.
Therefore, in case a model shows convergence problems (which can be indicated by warnings or problematic parameter values as mentioned in the quote) a good first strategy to address these problems is to remove the correlation among randomeffects parameters. We do this because these are the most difficult to estimate and they are often not of primary interest. Also removing them, often comes with comparatively little cost in terms of model precision.
More details and some really instructive examples of this are given in the “Parsimonious Mixed Models” paper by Bates, Kliegl, Vasishth and Baayen (2015): https://arxiv.org/abs/1506.04967
See especially Figures 3 and 5.There is also some relevant discussion on stats.stackexchange: https://stats.stackexchange.com/a/323341/442


AuthorPosts
 You must be logged in to reply to this topic.