Mercurywolf wrote: Based on your statement of 2 contributors providing 3/4 of the data you gleaned, this does not provide the best statistical analysis of Random Distribution. A better sample would be say 50 people reporting what they pulled for the exact same number of pulls, say 50-100. With 2 people providing data on 327 (75% of 436) pulls, your statistical analysis is off because this doesn't reflect the whole. As has been stated, the 10% rate of 5* is NOT per person, but an average of a whole. Unfortunately, if the two major contributors were facing less than a 10% distribution, your data is skewed because of the weight of the contribution. In a case like this, more accurate data is gleaned from having a larger sample size in the form of contributors, with each person contributing the same amount of data. Ex. If I were taking a survey, I would not be able to get good results if I gave 20 people a survey with 3 questions, and a different 5 people a survey with 20 questions. To get the best data, I'd have to give all 25 people the same survey with the exact same questions on each survey.
Mercurywolf wrote: Based on your statement of 2 contributors providing 3/4 of the data you gleaned, this does not provide the best statistical analysis of Random Distribution.
mgallop wrote: A quick follow up to Dave's excellent point. If you want to test that the distribution of 4s is off, you can't just check that each count is within the expected range, and put some stars next to it if p < .05, since if the stats are true, then we should expect 1 out of every 20 4s to have a "significant deviation" and there are 26 4s. My hunch is that the right way to do this would be a chi-squared test, where we see if the overall null ".33 for the 5s, .35 for the 4s" is rejected by the data.
Malenkov wrote: This is definitely an excellent point too; picking and choosing like I've done doesn't seem proper. But I would point out that while we might expect one out of 20 would be close to α, in this case we have 5 out of 27 at that point. Not as interesting as I thought, but maybe still a little interesting? In any case, on some of these dimensions we are close enough to the line that user error might easily account for statistical artifacts (see: "X-23 black").