Posting this out of curiosity, as I have once again gotten caught up with being overly "thorough" in my data analysis to the point that I can't seem to find a way to get past it and progress to actual analysis. This was spurred by seeing distributions that are not even remotely normally distributed on some widely used self-report measures in a quite typical protocol. After a thorough literature search I was unable to find ANYONE that did anything to address normality and I find it extremely unlikely this study was the first one where we saw a non-normal distribution on a measure that has been used hundreds, if not thousands, of times.
My colleagues seem convinced I obsess over these statistical issues when most people do not even look at these issues, so I guess I'm looking for some empirical evidence to either justify my (likely ridiculous) behavior or convince me I'm far more obsessive than is healthy. For the record, I do each of the poll options, will typically run a Box-Cox and/or do anywhere from 2-8 additional transformations (along with inclusion/exclusion of outliers) and will likely run any given analysis on several different versions to examine convergent validity of the outcomes and to test the limits of the data. I'm increasingly convinced most people just plug in the primary variable, maybe look for weird values if they don't like the results, and call it a day. Not sure I'm ever going to be comfortable with the latter, but could probably stand to find more of a middle ground.
Including a poll since I figured some may not want to admit publicly if the answer is that they don't check (though that also perhaps shows my bias
), but would also love to discuss the merits and costs/benefits of it. This is where I typically get caught up when writing. Once I have outcomes I'm confident in, I write fairly quickly, but can spend eons cleaning the data and getting it ready for analysis.
My colleagues seem convinced I obsess over these statistical issues when most people do not even look at these issues, so I guess I'm looking for some empirical evidence to either justify my (likely ridiculous) behavior or convince me I'm far more obsessive than is healthy. For the record, I do each of the poll options, will typically run a Box-Cox and/or do anywhere from 2-8 additional transformations (along with inclusion/exclusion of outliers) and will likely run any given analysis on several different versions to examine convergent validity of the outcomes and to test the limits of the data. I'm increasingly convinced most people just plug in the primary variable, maybe look for weird values if they don't like the results, and call it a day. Not sure I'm ever going to be comfortable with the latter, but could probably stand to find more of a middle ground.
Including a poll since I figured some may not want to admit publicly if the answer is that they don't check (though that also perhaps shows my bias

Last edited: