categorical variable

Member Training: Latent Class Analysis

August 7th, 2015 by

Latent Class Analysis is a method for finding and measuring unobserved latent subgroups in a population based on responses to a set of observed categorical variables.

This webinar will present an overview and an example of how latent class analysis works to find subgroups, how to interpret the output, the steps involved in running it.  We will discuss extensions and uses of the latent classes in other analyses and similarities and differences with related techniques.


Note: This training is an exclusive benefit to members of the Statistically Speaking Membership Program and part of the Stat’s Amore Trainings Series. Each Stat’s Amore Training is approximately 90 minutes long.

(more…)


Member Training: Dummy and Effect Coding

August 1st, 2014 by

Why does ANOVA give main effects in the presence of interactions, but Regression gives marginal effects?Stage 2

What are the advantages and disadvantages of dummy coding and effect coding? When does it make sense to use one or the other?

How does each one work, really?

In this webinar, we’re going to go step-by-step through a few examples of how dummy and effect coding each tell you different information about the effects of categorical variables, and therefore which one you want in each situation.


Note: This training is an exclusive benefit to members of the Statistically Speaking Membership Program and part of the Stat’s Amore Trainings Series. Each Stat’s Amore Training is approximately 90 minutes long.

Not a Member? Join!

About the Instructor

Karen Grace-Martin helps statistics practitioners gain an intuitive understanding of how statistics is applied to real data in research studies.

She has guided and trained researchers through their statistical analysis for over 15 years as a statistical consultant at Cornell University and through The Analysis Factor. She has master’s degrees in both applied statistics and social psychology and is an expert in SPSS and SAS.

Not a Member Yet?
It’s never too early to set yourself up for successful analysis with support and training from expert statisticians.

Just head over and sign up for Statistically Speaking.

You'll get access to this training webinar, 130+ other stats trainings, a pathway to work through the trainings that you need — plus the expert guidance you need to build statistical skill with live Q&A sessions and an ask-a-mentor forum.


Confusing Statistical Term #6: Factor

April 27th, 2012 by

Factor is confusing much in the same way as hierarchical and beta, because it too has different meanings in different contexts.  Factor might be a little worse, though, because its meanings are related.

In both meanings, a factor is a variable.  But a factor has a completely different meaning and implications for use in two different contexts. (more…)


6 Types of Dependent Variables that will Never Meet the Linear Model Normality Assumption

September 17th, 2009 by

The assumptions of normality and constant variance in a linear model (both OLS regression and ANOVA) are quite robust to departures.  That means that even if the assumptions aren’t met perfectly, the resulting p-values will still be reasonable estimates.

But you need to check the assumptions anyway, because some departures are so far off that the p-values become inaccurate.  And in many cases there are remedial measures you can take to turn non-normal residuals into normal ones.

But sometimes you can’t.

Sometimes it’s because the dependent variable just isn’t appropriate for a linear model.  The (more…)


Beyond Median Splits: Meaningful Cut Points

June 26th, 2009 by

I’ve talked a bit about the arbitrary nature of median splits and all the information they just throw away.Stage 2

But I have found that as a data analyst, it is incredibly freeing to be able to choose whether to make a variable continuous or categorical and to make the switch easily.  Essentially, this means you need to be (more…)


Multiple Imputation of Categorical Variables

June 1st, 2009 by

Most Multiple Imputation methods assume multivariate normality, so a common question is how to impute missing values from categorical variables.

Paul Allison, one of my favorite authors of statistical information for researchers, did a study that showed that the most common method actually gives worse results that listwise deletion.  (Did I mention I’ve used it myself?) (more…)