Just to dot some i's and cross some t's, a couple of points: (1) As far as I can tell, it is unclear why the "normal" distribution is called the normal distribution. Wikipedia has an entry on the normal distribution (yadda-yadda) and, in the history section, it is pointed out that de Moivre in 1783 first suggested the possibility of what we call the normal distribution. It was Gauss (1809) and Laplace (1774) who would provide a basis for thinking about the normal distribution as a probability distribution and this is why "really old school" statisticians refer to the normal distribution as a Gaussian distribution or Laplace-Gaussian distribution. See: http://en.wikipedia.org/wiki/Normal_distribution#Development
(2) The Wikipedia entry also provides some of the history for the different names that have been used for this distribution. Quoting from the entry: |Naming | |Since its introduction, the normal distribution has been known |by many different names: the law of error, the law of facility of |errors, Laplace’s second law, Gaussian law, etc. By the end of |the 19th century some authors[nb 6] had started using the name |normal distribution, where the word “normal” was used as an |adjective — the term was derived from the fact that this distribution |was seen as typical, common, normal. Peirce (one of those authors) |once defined “normal” thus: “...the ‘normal’ is not the average |(or any other kind of mean) of what actually occurs, but of what |would, in the long run, occur under certain circumstances.”[49] |Around the turn of the 20th century Pearson popularized the term |normal as a designation for this distribution.[50] http://en.wikipedia.org/wiki/Normal_distribution#Naming Why it should continue to be called the "normal distribution" is a puzzlement especially since we know that there are many different probability distributions and these may describe behavioral and/or psychological variables better than the normal/guassian/whatever. (3) I think that the use of the "normal distribution" in psychological testing, especially for IQ and other intelligence tests, promotes the use of the word "normal" because it is expected that "normal" intelligence is variable but extreme values in either direction are rare. So, 2 standard deviations above and below the mean in the normal distribution contain slightly more than 95% of the values in the distribution (thus "normal" or commonly expected) but values beyond this range have less than a 5% chance of occurring by chance (thus, "exceptional" or "special" or whatever euphemism we're using today to refer to people with intelligence way below "normal" or way above "normal"). Similar points can be made for other psychological variables (such as continuous measures of psychopathology). (4) Somebody should mention that what John Kulig and Jim Clark allude to below relates to the "central limit theorem". The Wikipedia entry on the normal distribution also coverts this topic; see: http://en.wikipedia.org/wiki/Normal_distribution#Central_limit_theorem and there is a separate entry on it (yadda-yadda) as well: http://en.wikipedia.org/wiki/Central_limit_theorem (5) I'm surprised no one asked Prof. Sylvester what he mean by "normal" or "normal distribution". It is possible that his definition is not consistent with common or "normal" definitions of those terms. -Mike Palij New York University m...@nyu.edu ------------ Original Message ----------- On Sun, 19 Jun 2011 21:12:41 -0700, Jim Clark wrote: Hi Here's a simple spss simulation of John's point about sums of multiple discrete factors being normally distributed. Just cut and paste into spss syntax window. input program. loop o = 1 to 1000. end case. end loop. end file. end input program. compute score = 0. do repeat v = v1 to v25. compu v = rv.uniform(0,1) > .5. end repeat. compute score1 = v1. compute score2 = sum(v1 to v2). compute score3 = sum(v1 to v3). compute score4 = sum(v1 to v4). compute score9 = sum(v1 to v9). compute score16 = sum(v1 to v16). compute score25 = sum(v1 to v25). freq score1 to score25 /form = notable /hist norm. Take care Jim James M. Clark Professor of Psychology 204-786-9757 204-774-4134 Fax j.cl...@uwinnipeg.ca >>> John Kulig <ku...@mail.plymouth.edu> 20-Jun-11 4:38 AM >>> Well, most things in psychology have numerous independent causes. Height is caused by at least several genes, your score on an exam is caused by answers to many individual questions, etc. The sum (i.e. adding) of independent events gets "normal" -- faster the more things you add (among other things). Example: Toss ONE coin, and record 0 for tails and 1 for heads. If this experiment - tossing one coin - is repeated long enough, you get about 50% heads and 50% tails (a flat or uniform distribution). Next toss two coins and record the total number of heads - it will either be 0, 1 or 2 heads. Repeat this experiment - two coins in one toss - and 25% of time you'll get 0 heads (TT) 50% of the time you'll get one head (since 1 head can be either TH or HT) and 25% of the time you'll get 2 heads (HH). With 0 1 and 2 heads on the X axis, it's not exactly a normal distribution but it is peaked at 1 head. When this is done by summing, say, number of heads when 5 or 6 coins are tossed in a single experiment, the resultant distribution (number of heads in one experiment) gets "normal" very fast (slower if the probability of a 'heads' is different than the probability of "tails" but it will still get to normal with enough coins in the experiment). Life is like coin tosses, no? Most everything we measure has multiple causes, so it should be no surprise that many things in the natural world are distributed "normally" .. though sometimes when a distribution has deviations from normality it's a clue about different underlying processes. IQ is somewhat normally distributed, though there is a little hump at the lower end (single gene effects?) and a slight bulge in the upper half (high IQ marrying other high IQ people?). Even when you measure the same exact thing over and over - like having all your students measure your height, their measurements will look normal .. classic psychological measurement theory would say that any measurement is the result of your "true" score added to an error component, and in many situations they assume "error" is unbiased, itself normally distributed, yaddy yaddy yaddy ... gets complicated quickly but the bottom line is that many things in the real world simply ARE normally distributed, or at least close enough to assume normality. A google search of the "central limit theorem" will give more precise information than this. On the other hand, I always tell my students to never take normality for granted, and merely LOOKING at data is the first step in determining if we can assume normality. Or at Yogi Berra put it, "you can observe a lot by looking" --- You are currently subscribed to tips as: arch...@jab.org. To unsubscribe click here: http://fsulist.frostburg.edu/u?id=13090.68da6e6e5325aa33287ff385b70df5d5&n=T&l=tips&o=11077 or send a blank email to leave-11077-13090.68da6e6e5325aa33287ff385b70df...@fsulist.frostburg.edu