Mixed Effect Models

From Sustainability Methods
Revision as of 09:03, 15 February 2021 by Christopher Franz (talk | contribs) (Created page with "450px|frameless|left|[[Sustainability Methods:About|Method categorization for Mixed Effect Models]] <br/> {|class="wikitable" style=...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Method categorization for Mixed Effect Models


Method categorization
Quantitative Qualitative
Inductive Deductive
Individual System Global
Past Present Future



not necessarily for every entry, but where it fits.

    • 1) background:**

Short: Mixed effect models are a statistical that allow for the integration of fixed effects as well as random effects.

Mixed effect models were a continuation of the introduction of random factors by Fisher into the Analysis of Variance. Fisher saw the necessity not only to focus on what we want to know in a statistical design, but also what information we likely want to minimise in terms of their impact on the result. Fsihers experiments on agricultural fields focussed to tame variance within experiments by replicates, yet he was strongly aware that underlying factors such as different agricultural fields and their unique combinations of environmental factors would necessarily impact the results. He thus developed random factor implementation, but Charles Roy Henderson took this to the next level by creating the necessary calculations to allow for linear unbiased estimates. These approaches allowed for the development of previously unmatched designs in terms of the complexity of hypotheses that could be tested, and also opened the door to the analysis of complex datasets that are beyond the sphere of purely deductively designed datasets. It is thus not surprising that mixed effect models rose to prominence in such diverse disciplines such as psychology, social science, physics and ecology. - when? - where? - by whom? - societal and scientific context in which the method originated - main topics and research communities in which the method is used

    • 2) What the method does**
    • categorization:**

-** inductive - deductive** - qualitative - **quantitative**

    • - individual - system - global**
    • - present-past-future**
    • how it works (description of the *how* and *why)***

Mixed effect models are mechanically speaking a step further in the combination of regression analysis and Analysis of variance, as they can combine the strength of the two by being able to incorporate both categorical and/or continous independent variables. Since mixed effect models were further developed into generalised linear mixed effect models, they are also able to incorporate different distributions concerning the dependent variable. Typically, the diverse algorithmic foundations to calculate mixed effect models such as maximum likelihood and restricted maximum likelihood allow for more statistical power, which is why mixed effect models often fair slightly better compared to standard regressions. The main strengths of mixed effect models is however not how they can deal with fixed effects, but that they are able to incorporate random effects as well. The total of all these possibilities make (generalised) mixed effect models the equivalent of the swiss army knife of univariate statistics. No statistical model to date is able to analyse a broader range of data, and mixed effect models are the gold standard in deductive designs. Often, these models serve as a baseline for the whole scheme of analysis, and thus these models already help research to design nd plane their whole data campaign. Naturally, these models are then the heart of the analyiss, and depending on the discipline the interpretation can range from widely established norms such as in parts of medicine to pushing the envelope in these parts of science where it is still unclear which variance is being tamed, and which cannot be tamed. Basically, all sorts of quantitative data can inform mixed effect models, and software such as R, Python and SARS offer broad arrays of analysis solutions, which arte still continuously further developed. Due to the complexity of the approach, mixed effect models see not easy to learn, as they are often hard to tame. Within such advanced statistical analysis, experience is key, and practice is essential in order to become versatile in the application of (generalised) mied effect models. - position in the research process: data gathering - analysis - interpretation - type of data - how and where do you gather the data? - which software / hardware does the method require? - What is necessary to know about the method?

    • 3) Strengths & challenges**

The biggest strengths of mixed effect models is how versatile they are. There is hardly any part of univariate statistics that can not be made by mixed effect models, or to rephrase it, there is hardly any part of advanced statistics that -even if it can be made by other models- cannot be made better by mixed effect models. They surpass Anova in terms of statistical power, eclipse regression by being better able to consider the complexities of real world datasets, and allow for a planing and understanding of random variance that brings science one step closer to acknowledge that there are things that we want to know, and things we do not want to know. Take the example of many studies in medicine that investigate how a certain drug works on people to cure a disease. To this end you want to know the effect the drug has on the prognosis of the patients. What you do not want to know if people are worse off if they are older, lack of exercise or have an unhealthy diet. All these single effects do not matter for you, because it is well known that prognosis often gets worse with higher age, and factors such as lack of exercise and unhealthy diet choices also influence the general health status and prognosis.What you may want to know if the drug works better or worse in people that have unhealthy diet choice, are older or lack regular exercise. These interaction can be meaningfully investigated by mixed effect models. All positive factor's variance is minimised, while the effect of the drug as well as it's interactions with the other factors can be tested. This makes mixed effect models su powerful, as you can implement them in a way that allows to investigate quite complex hypothesis or questions. The greatest disadvantage of mixed effect models is the level of experience that is necessary to implement them in a meaningful way.

Designing studies takes a lot of experience, and the current form of peer review does often not allow to present the complex thinking that goes into the design of advanced studies (Paper BEF China design). There is hence a discrepancy in how people implement studies, and how other researchers can understand and emulate these approaches. Mixed effect models re also an example where textbooks knowledge is not saturated yet, hence books are rather quickly outdated, and also often do not offer exhausting examples to real life problems resetter may fade when designing studies. Medicien and psychology are offering growing resources to this end, since here the preregistration of studies due to the reproducibility crisis offers a glimpse in the design of scientific studies. The lack of experience in how to design and conduct mixed feet mode driven studies leads to the critical reality that more often than not, there are flaws in the application of the method. While this got gradually less worse over time, it is still a matter of debate whether every published study with these models does justice to the original idea. Especially around the millennium there was almost a hype in some branches of science regarding how fancy mixed effect models were considered, and not all applications were sound and necessary. Mixed effect models can also make the world more complicated than it is, sometimes a regression is just a regression is just a regression.


    • 4) Normativity**

Mixed effect models are the gold standard when it comes to reducing complexities into constructs, for better of worse. All variables that go into a mixed effect model are normative choices, and these choices matter deeply. First of all, many people struggle which variables are about fixed variance, and which variables are relevant as random variance. Second, how are these variables constructed, are they continous or categorical, and if the latter, what is the reasoning behind the category levels? Designing mixed effect model studies is thus defiantly part of advanced statistics, and the is even harder when it comes to integrating non-designed datasets into a mixed effect model framework. Care and experience are needed to evaluate sample sizes, variance across levels and variables, which brings us to the most crucial point: Moodle inspection. Mixed effect models are build on a litany of preconditions, most of which most researchers choose to conveniently ignore. Now comes the real whammy: In my experience, this is more often than not ok, because it does not matter. Mixed effect models are -bless the maximum likelihood estimation- sturdy as hell. It is hard to find a model that is not having some predictive or explanatory value, even if hardly any pre-conditions are met. Does this mean that w should ignore these? No!! In order to sleep sound and save at night, I am almost obsessed with model inspection, checking variance across level, looking at the residuals, and looking for gaps and flaws in the models fit. We should be really conservative to this end, I think, because mind you that by focussing on fixed and random variance, we potentiate things that could go wrong. As I said, more often than not- this is not the case, but propose to be super conservative when it comes to you model outcome. In order to get there we need yet another thing: Model simplification. Mixed effect models lead the forefront of statistics, and this might be the reason why the implementation of AIC as a parsimony based evaluation criteria is more abundant here as compared to other statistical approaches. P-values fell widely out of fashion in many branches of science, as did the reporting of full models. Instead, models reduction based on information criteria approaches is on the rise, reporting parsimonious models that honour Occam's razor. Starting with the maximum model, these approaches reduce the model until it is the minimum adequate model in other words, the model that is as simple as possible, but as complex as necessary. The AIC is about the equivalent of a p-value of 0.12, depending on sample size, hence beware that the main question may be the difference from the null model. This links to the next point: Explanatory power. There has been some sort of a revival r2 values lately, mainly based on the suggestion of r2 values that can be utilised for mixed effect models. I deeply reject these approaches. First of all, mixed effect models are not about how much the model explains, but whether the results are meaningfully different from the null model. I can understand that in a cancer study I would want to know how much my model may help people, hence an occasional glance of the fitted value against the original values may do no harm. However, r2 in mixed effect models is to me a step into the bad old days when we evaluated the worth of a model because of its explained variance. This led to a lot of feeble discussions, of which I only mention here how good a model needs to be in terms of these values to be not bad, and vice versa. This is obviously a problem, and such normative judgements are a reason why statistics have such a bad reputation. Second, people are starting again to actually report their models based on the r2 value, and even have their model selection not be independent of the r2 value. This is something that should be bygone, yet it is not. Beware of the r2 value, it is only deceiving you in mixed effect models. Third, r2 values in mixed effect models are deeply problematic because they cannot take the complexity of the random variance into account. Hence r2 values in mixed effect models make us go back to the other good old days, and this is the days when mean values were ruling the outcomes of science. Today, we are closer to an understanding where variance matters, and why would we embrace that. Ok, it comes with a longer learning curve, but I think that the good old reduction to the mean was nothing but mean, sorry! Another very important point of mixed effect models is that they probably more than any statistical method remark the point where experts in one method -say for example interview- now needed to learn how to make interviews as a scientific method, but also needed to learn advanced statistics in the form of mixed effect models. This creates a double burden, and while learning several methods can be good to embrace a more diverse understanding, it is also challenging, and highlight a new development. Today, statistical analysis are increasingly outsourced to experts, and I consider this to be a generally good development. In my own experience, it takes a few thousand hours to become versatile at mixed effect models, and modern science is increasingly build on collaboration.


    • 5) Outlook**

In terms of mixed effect models, language barriers and the norms of specific disciplines are rather strong, and explaining the basics of these advanced statistics to colleagues is an art in itself, just as the experience of researchers being versatile in interviews cannot be reduced into a few hours of learning. Education in science needs to tackle this head on, and stop teaching statistics that are outdated and hardly published. I suggest that at least on a Masters level in the long run all students from the quantitative domain should be able to understand the preconditions and benefits of mixed effect models, but this is something for the distant future. Today. PhD students being versatile in mixed effect models are still outliers. Let us all hope that this statement will be outdated rather sooner than later. Mixed effect models are surely powerful and quite adaptable, and are increasingly becoming a part of normal science. Honouring the complexity of the world while still deriving value statements based on statistical analysis has never been more advanced on a broader scale, yet statisticians need to reecognize the limitations of real world data, and researchers utilising need to honour the preconditions and pitfalls of these analyses. Current science is in my perception far away from reporting reproducible analysis, meaning that one and the same dataset will be differently analysed by mixed effect model approaches, partly based on experience, partly based on differences between disciplines, and probably also because of many other factors. Mixed effect models need to be consolidated and unified, which would make normals science probably better than ever, I think.

Key Publications



The author of this entry is Henrik von Wehrden.