Good graphs are extremely powerful tools for communicating quantitative information clearly and accurately.
Unfortunately, many of the graphs we see today confuse, mislead, or deceive the reader.
Good graphs are extremely powerful tools for communicating quantitative information clearly and accurately.
Unfortunately, many of the graphs we see today confuse, mislead, or deceive the reader.
Multicollinearity is one of those terms in statistics that is often defined in one of two ways:
1. Very mathematical terms that make no sense — I mean, what is a linear combination anyway?
2. Completely oversimplified in order to avoid the mathematical terms — it’s a high correlation, right?
So what is it really? In English?
How do you know your variables are measuring what you think they are? And how do you know they’re doing it well?
Whether or not you run experiments, there are elements of experimental design that affect how you need to analyze many types of studies.
The most fundamental of these are replication, randomization, and blocking. These key design elements come up in studies under all sorts of names: trials, replicates, multi-level nesting, repeated measures. Any data set that requires mixed or multilevel models has some of these design elements. (more…)
Sample size estimates are one of those data analysis tasks that look straightforward, but once you try to do one, make you want to bang your head against the computer in frustration. Or, maybe that’s just me.
Regardless of how they make you feel, they are super important to do for your study before you collect the data.
After all, with the typical Type I error rate of 5% used in most tests, we are allowing ourselves to “get lucky” 1 in 20 times for each test. When you figure out the probability of Type I error across all the tests, that probability skyrockets.
(more…)