Re: Applied analysis question

2002-03-03 Thread Rich Ulrich


On 28 Feb 2002 07:37:16 -0800, [EMAIL PROTECTED] (Brad Anderson)
wrote:
> Rich Ulrich <[EMAIL PROTECTED]> wrote in message 
>news:<[EMAIL PROTECTED]>...
> > On 27 Feb 2002 11:59:53 -0800, [EMAIL PROTECTED] (Brad Anderson)
> > wrote:
BA > > > 
> > > I have a continuous response variable that ranges from 0 to 750.  I
> > > only have 90 observations and 26 are at the lower limit of 0, which is
> > > the modal category.  The mean is about 60 and the median is 3; the
> > > distribution is highly skewed, extremely kurtotic, etc.  Obviously,
> > > none of the power transformations are especially useful.  The product
> > 
[ snip, my own earlier comments ]
BA >
> I should have been more precise.  It's technically a count variable
> representing the number of times respondents report using dirty
> needles/syringes after someone else had used them during the past 90
> days.  Subjects were first asked to report the number of days they had
> injected drugs, then the average number of times they injected on
> injection days, and finally, on how many of those total times they had
> used dirty needles/syringes.  All of the subjects are injection drug
> users, but not all use dirty needles.  The reliability of reports near
> 0 is likely much better than the reliability of estimates near 750. 
> Indeed, substantively, the difference between a 0 and 1 is much more
> significant than the difference between a 749 and a 750--0 represents
> no risk, 1 represents at least some risk, and high values--regardless
> of the precision, represent high risk.

Okay, here is a break for some comment by me.

There are two immediate aims of analyses:  to show that
results are extreme enough that they don't happen by 
chance - statistical testing;  and to characterize the results 
so that people can understand them - estimation.

When the mean is 60 and the median is 3, giving report 
on averages, as if they were reports on central tendencies,
 is not going to help much with either aim.  If you 
want to look at outcomes, you make groups (as you did)
that seem somewhat homogeneous.  0 (if it is). 1.  2-3
eventually, your top group of 90+, which comes out to
'daily',  seems reasonable as a top-end.  Using groups 
ought to give you a robust test, whatever you are testing,
unless those distinctions between 10 and 500 needle-sticks
become important.  Using groups also lets you inspect, 
in particular, the means for 0, 1, 2 and 3.

I started thinking that the dimension is something like 
'promiscuous use of dirty needles';  and I realized that
an analogy to risky sex was not far wrong.  Or, at any rate,
doesn't seem far wrong to me.  But your  measure 
(the one that you mention, anyway) does not distinguish
between 1 act each with 100 risky partners, and 100 acts 
with one. 

Anyway, one way to describe the groups is to have some
experts place the reports of behaviors into 'risk-groups'.
Or assign the risk scores.   Assuming that those scores do
describe your sample, without great non-normality, you 
should be able to use averages of risk-scores for a technical
level of testing and reporting, and convert them back to the
verbal anchor-descriptions in order to explain what they mean.


[ ...Q about zero; kurtosis.]
RU > >
> > Categorizing the values into a few categories labeled, 
> > "none, almost none, "  is one way to convert your scores.  
> > If those labels do make sense.
> 
> Makes sense at the low end 0 risk.  And at the high end I used 90+
> representing using a dirty needle/syringe once a day or more often. 
> The 2 middle categories were pretty arbitrary.

[ snip, other procedures ]

> One of the other posters asked about the appropriate error term--I
> guess that lies at the heart of my inquiry.  I have no idea what the
> appropriate error term would be, and to best model such data.  I often
> deal with similar response variables that have distributions in which
> observations are clustered at 1 or both ends of the continuum.  In
> most cases, these distributions are not even approximately unimodal
> and a bit skewed--variables for which normalizing power
> transformations make sense.  Additionally, these typically aren't
> outcomes that could be thought of as being generated by a gaussian
> process.

Can you describe them usefully?  What is the shape of
the behaviors that you observe or expect, corresponding to
the drop-off of density near either extreme?

> In some cases I think it makes sense to consider poisson and
> generalizations of poisson processes although there is clearly much
> greater between subject heterogeneity than assumed by a poisson
> process.  I estimated poission and negative binomial regression
> models--there was compelling evidence t

Re: Applied analysis question

2002-03-03 Thread Rich Ulrich

On 27 Feb 2002 17:16:26 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

> i thought of a related data situation ...but at the opposite end
> what if you were interested in the relationship between the time it takes 
> students to take a test AND their test score
> 
> so, you have maybe 35 students in your 1 hour class that starts at 9AM ...
> 
> you decide to note (by your watch) the time they turn in the test ... and 
> about 9:20 the first person turns it in ... then 9:35 the second  9:45 
> the 3rd  9:47 the 4th ... and then, as you get to 10, when the time 
> limit is up ... the rest sort of come up to the desk at the same time
> 
> for about 1/2 of the students, you can pretty accurately write down the 
> time ... but, as it gets closer to the time limit, you have more of a 
> (literal) rush  and, at the end ... you probably put down the same time 
> on the last 8 students
> 
> you could decide just to put the order of the answer sheet as it sits in 
> the pile ... or, you might collapse the set to 3 groupings ... quick turner 
> iners, middle time turner iners ... and slow turner iners BUT, this clouds 
> the data
[ snip, rest]

Looks to me like it might be reasonable to re-sort and re-score
the speed as reciprocal, "questions per hour" -- instead of 
the original, hours per question.  That emphasizes something 
you (perhaps) omitted:  some tests at the end were incomplete.

Also, Q/H  accommodates that early test that was nearly blank.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Robust regression

2002-03-01 Thread Rich Ulrich

On 1 Mar 2002 00:36:01 -0800, [EMAIL PROTECTED] (Alex Yu)
wrote:

> 
> I know that robust regression can downweight outliers. Should someone
> apply robust regression when the data have skewed distributions but do not
> have outliers? Regression assumptions require normality of residuals, but
> not the normality of raw scores. So does it help at all to use robust
> regression in this situation. Any help will be appreciated. 

Go ahead and do it if you want.  

If someone asks (or even if they don't), you can tell 
them that robust regression gives exactly the same result.


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: help on factor analysis/non-normality

2002-03-01 Thread Rich Ulrich

On 1 Mar 2002 04:51:42 -0800, [EMAIL PROTECTED] (Mobile Survey)
wrote:

> What do i do if I need to run a factor analysis and have non-normal
> distribution for some of the items (indicators)? Does Principal
> component analysis require the normality assumption. 

There is no problem of non-normality, except that it *implies*
that decomposition  *might*  not give simple structures.
Complications are more likely when covariances are high.

What did you read, that you are trying to respond to?

>  Can I use GLS to
> extract the factors and get over the problem of non-normality. Please
> do give references if you are replying.
> Thanks.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Applied analysis question

2002-02-28 Thread Rich Ulrich

On 27 Feb 2002 14:14:44 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

> At 04:11 PM 2/27/02 -0500, Rich Ulrich wrote:
> 
> >Categorizing the values into a few categories labeled,
> >"none, almost none, "  is one way to convert your scores.
> >If those labels do make sense.

> well, if 750 has the same numerical sort of meaning as 0 (unit wise) ... in 
> terms of what is being measured then i would personally not think so SINCE, 
> the categories above 0 will encompass very wide ranges of possible values
[ ... ]

Frankly, the question is about meaning of numbers, 
and I would to ask it.

I don't expect a bunch of zeros, with 3 as median, and 
values up to 750.  Numbers like that *might*  reflect,
say, the amount of gold detected in some assays.  
Then, you want to know the handful of locations with 
numbers near 750.  If any of the numbers at all are big
enough to be interesting.

Data like those are  *not*  apt to be congenial for taking means.  
And if 750 is meaningful, using ranks is apt to be nonsensical, too.


In this example, the median was 3.
Does *that*  represent a useful interval from 0?  - if so, *that* 
tells me scaling or scoring is probably not  well-chosen.

Is there a large range of 'meaning'  between 0 and non-zero?  
Is there a range of meaning concealed within zero?
"Zero children" as outcome of a marriage can reflect 
(a) a question being asked too early; 
(b) unfortunate happenstance; or 
(c) personal choice
 - categories, within 0, and none of them are necessarily
a good 'interval'  from the 1, 2, 3...  answers.  But that 
(further) depends on what questions are being asked.


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Applied analysis question

2002-02-27 Thread Rich Ulrich

On 27 Feb 2002 11:59:53 -0800, [EMAIL PROTECTED] (Brad Anderson)
wrote:

> I have a continuous response variable that ranges from 0 to 750.  I
> only have 90 observations and 26 are at the lower limit of 0, which is
> the modal category.  The mean is about 60 and the median is 3; the
> distribution is highly skewed, extremely kurtotic, etc.  Obviously,
> none of the power transformations are especially useful.  The product

I guess it is 'continuous'  except for having 26 ties at 0.  
I have to wonder how that set of scores arose, and also, 
what should a person guess about the *error*  associated
with those:   Are the numbers near 750  measured with
as much accuracy as the numbers near 3?

How do zero scores arise?  Is this truncation;  the limit of
practical measurement;  or just what?

"Extremely kurtotic," you say.  That huge lump at 0 and skew
is not consistent with what I think of as kurtosis, but I guess
I have not paid attention to kurtosis at all, once I know that
skewness is extraordinary.

Categorizing the values into a few categories labeled, 
"none, almost none, "  is one way to convert your scores.  
If those labels do make sense.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Means of semantic differential scales

2002-02-26 Thread Rich Ulrich

> 
> > 2. Perhaps more likely, your boss may have learned
> > (wrongly?) that parametric stats should not be done unless scales
> > of measurement are at least interval in quality.
> 
> I don't know if his objection was to parametric statistics per se, but he did
> object to calculating means on these data, which he believes are only ordinal.
> 
> > Search on google
> > for people like John? Gaito and S.S. Stevens and for phrases like
> > "scales of measurement" and "parametric statistics."
> 
> Thanks.  Will do.
> 

Or,  do an Advanced search with  groups.google  
among the  sci.stat.*   groups for < Stevens, measurement >.
I think that would find earlier discussions and some references.
As I recall it, no one who pretended to know much would have
sided with your boss.

The firmness of Stevens's  categories was strongly challenged 
by the early 1950s.  In particular, there was Frederick Lord's 
ridiculing parable of the football jerseys.   (Naturally, psychology
departments taught the subject otherwise, for quite a while longer.)

Conover, et al., took a lot of the glory out of 'nonparametric tests'
by showing that you can't gain much from rank-order 
transformations, compared to any decent scaling.  That was 
in an article of 1980 or thereabouts.

I may have seen a 'research manual' dated as recent as 1985
that still  favored using rank-statistics with Likert-scaled items.  
I am curious as to what more recent endorsements might exist,  
in any textbooks at all, or in papers by statisticians.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: REQ: Appendix A. of Radford Neal thesis: "Bayesian Learning for Neural Networks"

2002-02-24 Thread Rich Ulrich

On Fri, 22 Feb 2002 18:00:16 +0100, Mark
<[EMAIL PROTECTED]> wrote:

> Hi,
> 
> I'm CS student interested in Radford Neal thesis called "Bayesian
> Learning for Neural Networks". I know that some years ago  this thesis
> was available for download from author's site, but nowadays there
> isn't possible. I have searched it on Intenet so I have not known to
> find it.

Why not send personal  e-mail and ask him?
He has posted to the stat-groups within the last month, from

 Radford Neal ([EMAIL PROTECTED])

> My e-mail address is [EMAIL PROTECTED] Please remove the
> "REMOVETHIS" string from the email address to get my real one. It's an
> anty-spam measure. I apologize for any inconvenience that it causes to
> you.

 - no inconvenience;  I won't bother.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: odds vs probabilities

2002-02-24 Thread Rich Ulrich

On 23 Feb 2002 06:49:58 -0800, [EMAIL PROTECTED] (Brad Branford)
wrote:

> hi,
> 
> thanks.  sorry if I posed the question poorly. actually, what I'm
> looking for is an intuitive understanding of when to use odds and when
> probabilities. I know that probs have a problem in that they don't
> make multiplicative sense: for instance, assume I have a probability
> of winning of 55%; if the likelihood of winning doubles, we have
> absurd outcomes if expressed in terms of probabilities.
> 
> thanks.

Uh-oh.  You have introduced a third technical term here.
Likelihood doesn't match either probability or odds.

There is a classical book by AWF Edwards by that title
(Likelihood), which argues we should be using likelihood 
for all our inference:  This makes some difference, though
I am not sure of what and when and how much.
Ps and likelihood compete in inference.  Ps and ORs arise
in descriptions of samples.

Where Probability is constructed as an area under the pdf
(often a 'tail area'), and OR is the ratio of two areas, 
the Likelihood is simply the height of the curve as evaluated 
at one point.  Thus  Probability and Odds are interchangeable, 
in just the way Ken describes.  

I think you only gain the 'intuitive understanding'  by 
exposure to a whole gamut of examples, including the 
counter-examples on when (especially) probability does 
not work because it is misleading.  For instance, 
in many contexts, 93% of a sample is  not "nearly the 
same as"  99%, since the OR is 7.00, and that will matter.

There is less reason to complain about  P  in place of ORs 
when the Ps are small -- where the arithmetic doesn't
expose the fallacy, as with your "twice 55%".   And P can
be approximately correct  in describing sample sizes when 
comparison-values are all between 25% and 75%, 
or 20-80.  But generally ORs are more appropriate for 
statistical models.  The drawback of ORs is that the public
is apt to understand them  less.

A year or so ago, someone posted journal references with 
arguments *against*  using ORs.  I looked up a couple
and did not find them impressive, but you can probably
find them by searching  sci.stat.*with  groups.google.com .

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluation of skating

2002-02-22 Thread Rich Ulrich

On 19 Feb 2002 15:14:01 -0800, [EMAIL PROTECTED] (Trevor Bond)
wrote:
[ snip, much ]
> affected who won the gold medal.  In fact, Looney (1994, p. 156) 
> concluded:
>   "all of the judges with an Eastern block or communistic background 
> not only ranked Baiul better than expected, but ranked Kerrigan 
> worse.  The same trend was seen for Western Block judges.  They 
> ranked Baiul worse and Kerrigan better than expected.  ... "

Finding a difference is one thing.  
Drawing invidious conclusions is a gratuitous step for
a statistician, isn't it?

Hypothesize.
Group A  holds a bunch of their own, inter-community skating
competitions.  So does group B.  This happens for many years.

I find it wholly reasonable -- if not expected -- that 
'community standards'  might exist with some 
divergence.  That's especially so when there were 
never any joint standards in the first place, and when
one country has the outstanding professional dance 
(ballet) of the world, which is accorded much local 
respect.  

>   When the 
> median of the expected ranks is determined, Kerrigan would be 
> declared the winner.  Before the free skate began, all the judges 
> knew the rank order of the skaters from the technical program and the 
> importance of the free skate performance in determining the gold 
> medal winner.  This may be why some judging bias was more prevalent 
> in the free skate than in the technical program."

Or, it could be (as the name suggests) that  'free skate'  offers 
more individual choices, more choices that will please or offend
personal tastes.

>   Looney's investigation of the effect of judge's ratings on 
> the final placement of skaters objectively validates what a chorus of 
> disbelieving armchair judges had suspected.  The median rank system 

Hooey.  You can't 'objectively'  validate one set of value-judgments.
You can't show that one set of scores arises 'by merit' while another,
with exactly the same salient features, does not.


The NY Times published the rankings in the pairs free program,
the one that ended up with ratings of the French judge being dropped.
There were 9 judges, labeled by nationality, and 20 teams.  
I don't know how the teams were 'qualified'  to appear here:  
there were  3 each, from Canada, Russia, and China.  In some
sense, anyway, these are the best in the world.  I have 
reproduced the data, below.

What astounds me is the uniformity of the rankings.  The *worst*
Pearson correlation between two judges (also, Spearman, 
since the scores are ranks) is 0.973, between judges from 
Japan and Russia.  Correlations with the total were above 0.98.

The NY Times highlighted the 'discrepancies' between each
judge and the Final ranking.  Of those 180 rankings, there were
two that were off by 3 (Japan rating the U.S. #13  as 10, for 
instance), 5 that were off by 2, and only 58 others  off by 1.

The most consistent rankings were by the French judge
(the scores that were thrown out).

Anyway, one consequence of that 'reliability'  is that there is 
relatively great 'statistical power'  for looking at blocs of votes,
if such exist.  I know some other rankings have been less
consistent than this; I don't know how (a)typical this level
of agreement might be for this skating event, or others.

Personally, I now suspect that there is 'collusion'  to the 
extent that judges agree, before the skate-off, about who
will be competing for 1-3 (say), 4-7, ...,  16-20.  
That might be decided on gross technical competence
(again, not invidious).
Concerns of great or small errors, difficulty, originality:  
these play a role within these strata.  And, biases about
tastes in presentations.

*= data: entered (for convenience) by judge.
* set up for SPSS to read; transpose; list; correlate.

Title   Skating Pairs, rankings by judge.
data list list / rank1 to rank20 judge(20F3.0,1x,A8).
begin data
  1  2  3  4  6  5  7  9  8 10 12 11 13 14 15 16 18 17 19 20 Russia
  1  2  3  5  4  7  6  8  9 10 11 13 12 15 14 16 17 18 19 20 China
  2  1  3  5  4  7  6  8  9 12 10 13 11 14 15 16 17 18 19 20 U.S.
  1  2  3  4  5  6  7  9  8 10 11 12 13 14 15 16 17 18 19 20 France
  1  2  3  4  5  6  7  8 10  9 11 12 14 13 15 16 17 18 19 20 Poland
  2  1  3  7  4  5  6  8 10  9 11 12 13 14 15 16 18 17 19 20 Canada
  1  2  3  4  5  6  7  8  9 10 11 12 15 14 13 16 17 19 18 20 Ukraine
  2  1  3  5  4  6  7  8  9 11 10 12 13 14 15 16 18 17 19 20 Germany
  2  1  3  4  5  7  6  8  9 12 11 13 10 15 14 16 17 19 18 20 Japan
end data.
execute.
flipnewnames= judge.
formats russia to japan(F2.0).
listall.
subtitle'Spearman' is the Pearson corr.
compute ranked= $casenum.
nonpar corr vars= russia to japan ranked

Re: How to test whether f(X,Y)=f(X)f(Y) is true??

2002-02-21 Thread Rich Ulrich

On Wed, 20 Feb 2002 22:21:38 -, "Chia C Chong"
<[EMAIL PROTECTED]> wrote:

[snip, various discussion before]
> 
> I have an example of data of 2 RVs. When I tested the correlation between
> them, by simply find the correlation coefficient, it shows that the
> correlation coefficient is so small and therefore, I could say that these
> two RVs are uncorrelated,or better still, not linearly correlated. 

Right!
>   However,
> when I plotted the scatter plot of them, it is clearly shown that one of the
> varriable does dependent on the other variable  in some kind of pattern, is
> just that there are not lineraly dependent, hence the almost zero
> correlation coeffiicent. So, I am just wonder whether any kind of tests that
> I could use to test dependency between 2 varaibles...

Construct a test that checks for features.  What features?
Well, what features characterize your *observed*  dependency,
in a generalized way?  -- you do want a description that would
presumably have a chance for describing some future set of
data.

The null hypothesis is that the joint density is merely 
the product of the separate densities.  
For a picture:  a greytone backdrop changes just gradually, 
as you move in any direction.  Distinct lines or blotches are
'dependencies' -- whenever they are more distinct
than would  'arise by chance.'

The best test to detect vague blotches would not be the
best to detect sharp spots, and that would be different
from detecting lines.

As I wrote before , 
> >
> > So there is an infinite variety of tests conceivable.
> > So the  *useful*  test is the one that avoids 'Bonferroni correction,"
> > because it is the one you perform because
> > you have some reason for it.
> >
--
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: How to test whether f(X,Y)=f(X)f(Y) is true??

2002-02-20 Thread Rich Ulrich

On Wed, 20 Feb 2002 19:30:19 -, "Chia C Chong"
<[EMAIL PROTECTED]> wrote:

> 
> "Vadim and Oxana Marmer" <[EMAIL PROTECTED]> wrote in message
> [EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
> > You can start with checking if they are correlated. It's simpler to do. If
> > you find that they are correlated then you have the answer to your
> > question.
> > If you find that they are uncorrelated and you have a reason to believe
> > that they may be not independent anyway then you can look for more
> > advanced tests.
> 
> Can you give some examples of more advanced tests that can be used to test
> the depedency of data when there these data are uncorrelated??

You can check for an obvious non-linear (say quadratic) fit.

WHAT  is your 'reason to believe that they may be 
not independent'?  
Anything that makes any pattern, at all, is 'dependent.'  

So there is an infinite variety of tests conceivable.  
So the  *useful*  test is the one that avoids 'Bonferroni correction,"
because it is the one you perform because 
you have some reason for it.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: SPC control limits

2002-02-20 Thread Rich Ulrich

On 20 Feb 2002 08:50:19 -0800, [EMAIL PROTECTED] (MrTequila)
wrote:

> Hi all, hope this is the right place. 
> 
> i was just wondering what you should do when you establish some
> control limits but some of the data points you've just used are
> outside of the limits you just established?
> 
> should you write them off as bad, leave them or go back and see/fix
> the problem?

 - looks like an early chance to test them.  If you go back
and there is *not*  any problem, then here's a hint that 
your limits may not be useful after all.

I suppose, if you  *know*  you did badly, you could write
yourself a failing grade -- based on guilt-feelings rather
than on data -- and change the limits without ever 
testing whether there is a real problem.  

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Numerical recipes in statistics ???

2002-02-19 Thread Rich Ulrich

On 19 Feb 2002 13:13:08 -0800, [EMAIL PROTECTED] (The Truth) wrote:

> Glen Barnett <[EMAIL PROTECTED]> wrote in message 
>news:<[EMAIL PROTECTED]>...

> > The Truth wrote:
> > > Are there any "Numerical Recipes" like textbook on statistics and probability ?
> > > Just wondering..
> > 
> > What do you mean, a book with algorithms for statistics and probability
> > or a handbook/cookbook list of techniques with some basic explanation?
[ ... ] 
> 
> I suppose I should have been more clear with my question. What I
> essentially require is a textbook which presents algorithms like Monte
> Carlo, Principal Component Analysis, Clustering methods,
> MANOVA/MANACOVA methods etc. and provides source code (in C , C++ or
> Fortran) or pseudocode together with short explanations of the
> algorithms.

Unreasonable dreams.

Maybe your university would have some trace of
the original BMD package.  I think it was in the public domain, 
and that it was the source of many routines for
SPSS, SAS, and BMDP.  But those were simple routines.

That should have Principal components, and perhaps some 
clustering.

However:  There are dozens of Monte Carlo methods;
you might find a book for whatever your particular *field*
happens to be.

And check the SPSS 'proximities'  routines to get a 
notion of the complexity of a driver-routine for clustering.
It has 3 dozen or so criteria for closeness.  Clusters can 
use [some criterion]  and apply it to each case as it
relates to each other case, or to each previous group.
Or it can relate created-groups to other created-groups.
Then there is the big choice of whether cases start
separate, to be joined;  or start out joined, to become 
separate.

Compared to Clustering, MANOVA might have some
pretty good routines.  But I have not seen much great
program organization for MANOVA procedures, 
once you get beyond 'discriminant function'  and 
the paradigm of multi-regression with multiple-criteria.

'Pseudocode,'  however, is almost what it looks like when
you start with matrix routines, in SPSS or  R  or 
various specialty programs.  You do have to speak the 
language of matrices, so you know when it is, that 
you want to get an eigenvector; and what to do with it.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: can multicollinearity force a correlation?

2002-02-19 Thread Rich Ulrich

On 18 Feb 2002 16:29:27 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> > You should take note that R^2  is *not*  a very good measure
> > of 'effect size.' 
> 
> Hi Rich, you asked to see my data, 

 - I don't remember doing that -

>i've posted the visual at the
> following location http://www.accessv.com/~joemende/insulin2.gif note
> that the r^2 is low despite the fact that it agrees with common sense:
>  Insulin levels are shown here to decrease with increasing exercise as
> well as with decreasing food intake..   My r^2 is low but i think it
> is clear that the above is true..
> 
> I've included several different views, "rating" is in MET values, i
> forgot to multiply against body weight in kg to get KCAL spent per
> day..

If the scattergram is meaningful, blue-ish and to the right, then 
the big, 3-D shaded bar-graph to the left may be thoroughly
misleading.  I assume that the plane imposed on both represents
the regression.

Do the dots in the scattergram mean something?  
I  make sense of them;  but then the bars seem to be a 
bad and misleading choice of plotting something-or-other.
The bars are an apparent picture of something greater
than  R^2  of 2% -- so I am sure they don't represent *that*
relationship in a proper way.  

My tentative conclusion is that your 2%  effect  really
is a small one; it should be difficult to discern among 
likely artifacts; and therefore, it is hardly worth mentioning

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: covariates !!

2002-02-19 Thread Rich Ulrich

On Mon, 18 Feb 2002 20:57:36 +0100, "jan plessers"
<[EMAIL PROTECTED]> wrote:

> Hello,
> 
> I did a likert on 2 groups, a Mann-whitney showed that ther was a
> significant difference between the 2 groups. 

If 'did a Likert'  means what I expect, and 
the scaling was decent enough to be worth commenting on, 
then the rank-transformation was unneeded and wasteful of information.

> The next thing I did were some
> univariate statisics between the covariates and I found some significant
> differences between the covariates (ex. age sign. differs between the 2
> groups).
> My questions are:
> Can I test again for a sign. difference between the 2 groups, but with the
> covariates controlled ? How to do this? How to do this in SPSS ?

If the groups aren't random, 
and especially when they are non-randomly non-random,
then you have to build the argument that covariates don't matter.

First, you can stick them in and see if anything changes.  
If not, then you are okay.
You do want to ask the procedure to interpret the covariates 
'before'  or  'at the same time as'  
(not, 'after')  the main effects.

I've always done this in SPSS  (6.1 and earlier) with 
ANOVA   vara  by grps(1,4) with covar/

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Newbie question

2002-02-17 Thread Rich Ulrich

On 15 Feb 2002 14:38:49 -0800, [EMAIL PROTECTED] (AP) wrote:

> Hi all:
> 
> I would appreciate your help in solving this question.
> 
> calculate the standard deviation of a sample where the mean and 
> standard deviation from the process are provided?
> E.g. Process mean = 150; standard deviation = 20. What is the SD for 
> a sample of 25?  The answer suggested is 4.0

Here is a vocabulary distinction.   Or error.
I don't know if you are repeating the problem wrong, or 
you are speaking from a tradition that I am not familiar with.

As I am familiar with it, statisticians say that 
"the standard deviation"  is the "standard deviation of the sample."

We say that the "standard deviation of the sample *mean*" 
will be frequently referred to as the "standard error";  and 
"The SD of the mean [or the SE] equals SD/sqrt(N)".

That is confusing enough.  
I hope this makes your sources clear.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: can multicollinearity force a correlation?

2002-02-15 Thread Rich Ulrich

[ snip, previous problem]
> 
> This is similar to a problem I have come across: the measurement of a
> serum value against exposure.
> My theory is that they are correlated.  But the data says that they
> have an R^2 of 0.02 even though the  p-value for the beta is p=1E-40 
> (ie. zero).
> 
> As you explain this is possible.  My reasoning is that the exposure is
> happening  many hours before the measurement of serum and so that is

You should take note that R^2  is *not*  a very good measure
of 'effect size.'  It only works when you are repeating something
familiar.  You have seen so-much before, and you may be 
happy to see as much again;  but it does not tell you as much 
as knowing that there is a 4-fold Odds Ratio for a factor -- 
which is the usual measure when you have a rare dichotomy, or
something that can be conveniently described that way.

[ snip ]
>   .  R^2 is very useful though, for example if
> you want to know in the american population what is the highest source
> of fat, you would use R^2 on the food frequencies, not the beta
> coefficient.. because the R^2 would tell you the food that most
> predicts, rather than the "strength" of the effect of the food..  ie.
> low fat foods may be main source of fat in diet..
> 
> -just thinking outloud hehe..

Well, maybe R^2  is useful.  But you need to know how it is anchored.
Do you have continuous variables? - I thought you had dichotomies,
where the Odds Ratio is rule, when you have small rates.
A 'coefficient of determination'  or R-squared of  0.18
reflects *at least*   a 4-fold increase in Odds Ratio when 
the 4 cells are all around 50% -- For that 0.18,  the
OR  is  higher, if the margins are less balanced.

And also.
The R-squared is going to describe the sample-on-hand:  
If you sample with too-narrow variation, you
get  R-squared that is too-small.  Similarly, for large.   
The beta describes co-variation in another way; the raw beta
(not the standardized)  is usually what is more interesting,
if you really have a large enough N   that the actual coefficients
are interesting (mine usually are not).

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: homogeneity of variance

2002-02-15 Thread Rich Ulrich

On 14 Feb 2002 17:14:24 -0800, [EMAIL PROTECTED] (Thomas Souers)
wrote:

TS>
> I have another question regarding one-way ANOVA. I have noticed 
> that in some books, nothing is said about what you can do when the
> factor level variances are unequal. In Neter's big book,
> transformations are recommended. "

Well, there are a lot of excuses for comparing *two*  means.
It is tougher to make up reasons why you ought to compare
more than two, when they are arbitrary sets of numbers that
are not well-behaved.  For instance, they are  *potentially* 
well behaved, in my book, if they simply happen to have been
recorded or measured in the wrong units/ scale -- so they
will be fixed by transformation.  You probably ought to be
re-thinking your whole scientific hypothesis and test, if 
your problem is worse than that.

TS>
> If the data are approximately normal, why not just use a 
> Satterthwaite approximate t-statistic for pairwise comparisons? 
> For example, you could use a Bonferroni type procedure. 
> What are people's opinions about this approach?"

You can see my comments (today)  on another thread, 
about t-tests.  If Ns are equal, it doesn't matter which test.
So you might as well use the regular apparatus.  If they 
are not, both tests are fairly rotten by one-tailed criteria.

By the way:  What clinical research uses in followup testing
for unequal N   is either  Bonferroni correction, or one or 
another  *approximation* (which may not be very good).
(I think I would never trust 'exact procedures' that may 
have been designed for unequal N, generically.)  

Personally, I try to do important tests and avoid Bonferroni.


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: If T-Test can not be applied

2002-02-15 Thread Rich Ulrich

On Thu, 14 Feb 2002 23:48:02 +0100, "Matthias" <[EMAIL PROTECTED]>
wrote:

> Hello,
> 
> would be nice if someone can give me some advice with regard to the
> following problem:
> 
> I would like to compare the means of two independent numerical sets of data
> whether they are significantly different from each other or not. One of the
> two underlying assumption to calculate the T-Test is not given (Variances
> are assumed to be NOT equally distributed; but data is normally
> distributed). What kind of (non?)parametric-test does exist - instead of the
> T-Test - to calculate possible differences in the two means?
 [ ... ]

The *logical* or measurement  problem raised by 'different 
variances'  is manifested when one sample predominates 
at both extremes.  In that case, a conclusion about 'superiority'
depends on someone's scaling or weighting of scores.  That 
problem does not exist when the difference between groups is 
a shift of the mean, or a stretching of the distribution to one side.

Comparing *means* ...  Are you sure that you want to focus on means?
Then it might as well be the t-test, evaluated one way or another:
equal variances-assumption;  unequal;  or  randomization.

Other than means:
The problems that lead to 'other tests'  are ones that say the means 
are 'meaningless', and that you want to look at some version of 
stochastic superiority.  But if the shapes are similar (same family),
then the t-test with equal Ns  is accurate; and with vastly unequal
Ns, the two standard tests are both  biased and 
potentially misleading -- so you have to look at both, and decide 
whether you should prefer one over the other based on (for instance) 
where and how the data were collected, or how they arise.

If the shapes of distributions are not similar, then
rank-transformation is not reliable or robust for 'superiority.'
You can compute some Robust tests, such as the 2x2  contingency
test, which may be based on median split or whatever split meets
your needs.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: one-way ANOVA question

2002-02-14 Thread Rich Ulrich

On 13 Feb 2002 09:48:41 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

> At 09:21 AM 2/13/02 -0600, Mike Granaas wrote:
> >On Fri, 8 Feb 2002, Thomas Souers wrote:
> > >
> > > 2) Secondly, are contrasts used primarily as planned comparisons? If 
> > so, why?
> > >
> >
> >I would second those who've already indicated that planned comparisons are
> >superior in answering theoretical questions and add a couple of comments:
> 
> another way to think about this issue is: what IF we never had ... nor will 
> in the future ... the overall omnibus F test?
> 
> would this help us or hurt us in the exploration of the 
> experimental/research questions of primary interest?

 - not having it available, even abstractly, 
would HURT, because we would be 
without that reminder of  'too many hypotheses'.

In practice, I *do*  consider the number of tests.
Just about always.

Now, I am not arguing that the particular form 
of having an ANOVA omnibus-test  is essential.
Bonferroni correction can do a lot of the same. It just
won't always be as efficient.

> i really don't see ANY case that it would hurt us ...
> and, i can't really think of cases where doing the overall F test helps us ...
> 

But, Dennis, I thought you told us before, 
you don't appreciate  "hypothesis testing" ...
I thought you could not think of cases where doing
*any*  F-test helps us.

[ ... ]

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Book/textbook on applied math with expectation/variance

2002-02-10 Thread Rich Ulrich

On Sat, 9 Feb 2002 01:17:14 +0900, "maximus" <[EMAIL PROTECTED]> wrote:

> It may seem odd with the question in the title, but I want to read and have
> some more
> practice with (applied) math with expectation/variance, which is in many
> forms, for example
> with max/min, integration (inside or outside the integrant, or on the
> limit), exponent, etc...
> 
> Not many books offer more than just the simple definitions and examples.
> 
> Is there any particular book that have more ?

Level of difficulty?  
Kendall and Stuart's volumes of "The Advanced Theory
of Statistics" could occupy you for a while. 
 
Or look in the same section of the library as K&S; 
or whatever sources you have liked before.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Ansari-Bradley dispersion test.

2002-02-10 Thread Rich Ulrich

On Sat, 09 Feb 2002 16:59:34 GMT, Johannes Fichtinger
<[EMAIL PROTECTED]> wrote:

> Dear NG!
> I have been searching for a description of the Ansari-Bradley dispersion test up to 
>now for 
> analysing a psychological research. I am searching for a description of this test, 
>specially a
> description how to use the test.
> 
> Please, can you tell me, how to use the test, or show me a link, where it is 
>described?
> Thank you very much in advance,

I plugged Ansari-Bradley  into a search by  www.google.com  and 
there were 287  hits.  The first page contained the (aptly named)

http://franz.stat.wisc.edu/~rossini/courses/intro-nonpar/text/Specifications_for_the_Ansari_Bradley_Test.html

I suggest repeating the search.  That also eliminates the "pasting"
problem if your reader has broken  the long URL into two lines.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: can multicollinearity force a correlation?

2002-02-07 Thread Rich Ulrich

On 5 Feb 2002 18:01:15 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> > You made a model with the "exact same exposure in different units",
> > which is something that no one would do, 
> 
> Hehe, translation is don't post messages until you've thought them
> through.
> 
> Anyway, turns out that the answer to my question is "No"..

Well, I think I speak for several statisticians when I say that
we still don't know what you refer to as 'multi collinearity'.  Do you
mean  100%, as in your question?   What *are*  you asking?

> Multicollinearity cannot force a correlation.  It turns out that ONE
> of the variables *was* correlated With R^2=0.45 and so
> multicollinearity had no effect on overall R^2.
 [ ... ]

If you are concluding that you won't improve R^2  by using
exactly the same variable twice, you are correct.  

Another post-er  has suggested where  you  *do*
have to watch out for multi-colliinearity:  He described the
case where the multiple R^2  is large despite small univariate 
correlations with the criterion.  (It was wordier than that, 
but that is what he did.)  

For further information-> You could search the last few months
of posts in sci.stat.*  usinggroups.google.com   and look for
'confounding'  or  'masking';  and there might be something 
more in my own stats-faq.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: EViews 3.1 error message

2002-02-07 Thread Rich Ulrich

On 6 Feb 2002 11:31:10 -0800, [EMAIL PROTECTED] (James McQueen) wrote:

> Hi,
> 
> I tried to run a workfile on EViews 3.1 that I had created in EViews
> 4.0. It is a three equation system estimated via full information
> maximum likelihood. It worked fine in 4.0. Now, I'm using 3.1 and it
> opens the 4.0 workfile no problem but when I hit the estimate button
> (via FIML) it returns: error, near singular matrix. As far as I know
> the workfile was unchanged from the one that worked fine in 4.0. Do I
> need to create a new workfile in 3.1 and re-enter the equations and
> re-import the excel file to make it work in 3.1?

You are trying to import the 4.0  version of something into its 
3.1 version?  that sounds like a formula for screwing up, if 
that package isn't smart enough to prohibit  it, or to tell you
the explicit limitations or methods.  Have you looked for 
"Export to old version"?

Sometimes packages can't read their own *old*  versions; 
and it sounds like you tried to read the *newer*  version with
an older one.  You can try re-importing just the equations 
(or just the data) separately, but if you have doubts, it might 
take more time to check the total accuracy  than it would 
to start from scratch.

Can you re-import your data from Excel?  That sounds safe,
even if you create it in Excel by exporting from 4.0.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Case-cross over

2002-02-07 Thread Rich Ulrich

On Thu, 7 Feb 2002 13:20:44 +0100, "Anna Axmon"
<[EMAIL PROTECTED]> wrote:

> Hi,
> 
> does anyone know if there is a textbook on case-cross over design?
> 

"case crossover design"  (in quotes) gets 258 hits reported in google.

I did not notice a textbook review, but those should lead you 
to whatever is available.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: time series analysis

2002-02-07 Thread Rich Ulrich

 - I don't see a public answer to this one 
from a couple of weeks ago -

On 25 Jan 2002 10:16:45 -0800, [EMAIL PROTECTED] (Jack Eslick)
wrote:

> I am working with a data that were collected at non-uniform time
> intervals.  I want to use regression analyses to determine if there is
> a temporal trend.  Common sense tells me that the data need to be
> assigned some type of weight value (data collected at closer intervals
> should have less ability to influence the regression line).  I have
> been unable to find a reference about how to assign the weights, other
> than some vague references to "an appropriate" method.  Can someone
> recommend an approachable reference?

On careful re-reading, this is not the question that I expected
from its first line.  

Now I gather that the precise lack of uniformity does not bother 
you for its lack of precision.  You are (possibly) quite satisfied
with what  you have as a time line, for whatever it says about
the passage of time.  But you have some individuals with many
extra (or fewer) points than others, or some parts of the time 
lines, by individual or for the whole sample, that are not 
equivalent in data density.  - That is fuzzy on my part, but I am 
not highly confident in (for instance) my basic reading,  that 
you have  'repeated measures.'

Why does common sense tell you that you need weights?

How extreme is the problem? - Perhaps you are underestimating
the robustness of regression.  What is the design (N and repeats)
and what is the approximate R-squared:  Are you trying to model 
these data on hand with great confidence, and place accurate
intervals on coefficients, or are you trying to squeeze out 
a statistical test that will be legitimate?

There is not always a unified solution.  For instance, I can
imagine that you might be best served by partitioning cases
(I am assuming repeated measures again)
into 'sparse data'  and 'heavy data'; analyzing separately;
and combining results.  You do have to explain, at some
time, why some data are sparse, and argue that it makes
no important difference, right?

I don't know of particular references for what I have been
talking about, but "unbalanced data" might help your search.  

And you could try us with more detail, if you wish.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Sensitivity Analysis

2002-02-06 Thread Rich Ulrich

On 31 Jan 2002 10:06:36 -0800, [EMAIL PROTECTED]
(Christopher J. Mecklin) wrote:

> I had a colleague (a biologist) ask me about sensitivity analysis.  I am 
> not familiar with the technique (above and beyond knowing that the 
> technique exists). What books/articles/websites/etc. would be good sources 
> for my colleague to learn about sensitivity analysis.  Since he's a 
> biologist and not a statistician, I'm assuming he would prefer a treatment 
> geared towards application rather than theory.

I have not seen any reply to this.  I suspect that there might be
too many options that refer to 'sensitivity'  and none of us
are sure what you are interested in, precisely.

What's another keyword?  I pair specificity with sensitivity;  but
I don't refer to  'sensitivity analysis', I say 'discriminability.'
Your question -- and my background thoughts of 1000-generation, 
simulation analyses in genetic model ling -- makes me think of
something I saw years ago, called  'perturbation analyses'.

Try Google, or try us again with additional detail.

Hope this helps.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Interpreting mutliple regression Beta is only way?

2002-02-05 Thread Rich Ulrich

On 4 Feb 2002 16:14:11 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> > 
> > In biostatistical studies, either version of beta is pretty worthless.
> > Generally speaking.
> 
> If I may be permitted to infer a reason:
> if you have  
> 
> bodyweight= -a(drug) - b(exercise) + food
> 
> Then the standardized coefficients will affect bodyweight but they
> will also affect each other.  They would only be useful if drug intake
> was perfectly independant of exercise and food in the population.

Okay, that is a bit of a reason, which applies widely;  but that 
is not what I was pointing at when I wrote that line.

I was over-stating.  But many of the results in clinical research
are 'barely statistically significant'  and when that is so, then
(in my opinion), the 95% Confidence Interval does not add much 
to the statement of "test result", and the point estimate 
of the effect (beta, or a mean-difference)  is pretty loose, too.

I want to beat a 5% alpha.  And I want a result to have a
50% interval (say)  that is actually interesting, and well-above
the measurement jitters. -- That is an *essential*  requirement
for being coherent, if you want to describe 
observational studies.  And it is a pretty good idea for
writing about randomized-controlled studies, too.

> 
> If they are not independant but partially collinear (0.5) using linear
> regression is it possible to know whether the drug is strong enough
> (colloquially speaking) to recommend?
> I assume that it would be impossible as a change in drug cannot be
> separated from a change in exercise in the population.  Ie. people are
> exercising and taking the drug so it is impossible to distinguish
> which one is beneficial.

You can look at the zero-order effect (raw correlation) as well as
the partial effect after controlling for other potential predictors:
one at a time, or in sets.  If your Drug always shows the same
prediction, regardless of how you test it, that is a pretty good sign.

> I've heard of "ridge regression" will try to investigate this area
> more..
> will probably figure it out with time hehe..

Ridge regression can be decomposed into a combination of the
p-variate regression, averaged with the 1-variable regressions, 
and all the i-variable regressions that lie in-between.  There is
not much gain using Ridge if you avoid suppressor relationships 
from the start -- all those cases where a beta is the opposite sign
from its zero-order beta.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: can multicollinearity force a correlation?

2002-02-05 Thread Rich Ulrich

On 5 Feb 2002 08:28:05 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> Is it possible that multicollinearity can force a correlation that
> does not exist?
> 
> I have a very large sample of n=5,000
> and have found that
> 
> disease= exposure + exposure + exposure + exposure R^2=0.45
> 
> where all 4 exposures are the exact same exposure in different units 
> like ug/dL or mg/dL or molar units.

Now, see, that is totally, thoroughly ignorant.

You made a model with the "exact same exposure in different units",
which is something that no one would do, who had understood 
even one single semester of  intro-to-statistics.

> 
> Nonetheless when I do a simple correlation (pearson) I found that the
> exposure in ug/dL did not affect the disease.
> 

I find it extremely likely that you do not know how to read
the computer printout and then tell us what you have read.

> This seems hard to believe as my sample is relatively large..
> I don't believe the 0.45 R^2 is possible but was shocked by it.  I'll
> try to rerun it in other, more realistic models.

My advice is:  Please ask for local help from someone 
who can lead you through analyses, step by step.

If you insist on asking like this, I suggest that you 
cut-and-paste some computer output;  or e-mail
the full dataset to your eventual e-mail  helper.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: The exam questlons will probably be tougher

2002-02-03 Thread Rich Ulrich

On 3 Feb 2002 03:34:35 -0800, [EMAIL PROTECTED] wrote:

[ snip, previous examples ... ]
> 
> Asked about statistics practice problems. Search took 0.35 seconds
> Got about 945,000. hits.
> 
> Nuff said,

 - Right idea, sloppy technique;  I would say in my critique.

Google hits 966,000 for me, for ,
but that means it found more than one of the three words, 
somewhere.  Probably.  Too many to scan much of.

When you get that sort of excess, you should choose 
to guide the search a little more, by  using Advanced
Search  (or using quotes to get the same thing).

There were only 67 for <"statistics practice problems">, 
some of them designed for a particular class's test.
 
Quick inspection shows that the latter two words should be
promising.   nets 7420, so
you might be able to select topics within those.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: R-square & parameter estimates

2002-02-01 Thread Rich Ulrich

On 31 Jan 2002 08:26:19 -0800, [EMAIL PROTECTED] (Yvette) wrote:

> Hi all,
> 
> My question, although probably basic to most of you, is:  If you are
> comparing two models, why might the test variables parameter estimates
> be significant in the second case and not in the first yet the
> R-square is decreased.  For example:
> [ snip, rest ]

"Comparing two models"  usually means something in
particular to statisticians:   there is an overall test.
That is not what you mention or describe.

I can construct for you a HIGHLY accurate model with several
variables, where none of the individual coefficients (which
measure 'partial'  effects)  will be important.
 - look up subjects like, "interpreting (partial) coefficients"
and discover how awkward that is.


Hope this helps.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Verba Volant 31-01-02

2002-01-31 Thread Rich Ulrich

On Thu, 31 Jan 2002 10:37:13 -0500, Rich Ulrich <[EMAIL PROTECTED]>
wrote:

> Postmaster,
> Would you please tell your User that they have to
> stop sending their daily message to Usenet Newsgroups.
> 
> Below are the headers, and the contents of what was 
> received today, at sci.stat.edu.
> 
> Rich Ulrich, [EMAIL PROTECTED]
> == headers and message (raw plus HTML plus encoded)

ooops, sorry folks, my apologies.

I think I sent that to the proper postmaster;  but
I thought I was *not*  sending it to the newsgroup.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Verba Volant 31-01-02

2002-01-31 Thread Rich Ulrich

Postmaster,
Would you please tell your User that they have to
stop sending their daily message to Usenet Newsgroups.

Below are the headers, and the contents of what was 
received today, at sci.stat.edu.

Rich Ulrich, [EMAIL PROTECTED]
== headers and message (raw plus HTML plus encoded)
Path:
pitt.edu!nntp.club.cc.cmu.edu!micro-heart-of-gold.mit.edu!enews.sgi.com!howland.erols.net!newsfeed.lotsanews.com.MISMATCH!pln-e!spln!dex!newsgate.newsguy.com!newsp.newsguy.com!mdrn
From: [EMAIL PROTECTED] (Verba Volant)
Newsgroups: sci.stat.edu
Subject: Verba Volant 31-01-02
Date: 31 Jan 2002 03:56:55 -0800
Organization: None
Lines: 293
Message-ID: <5.1.0.14.2.2002013927.00a4a210@localhost>
NNTP-Posting-Host: mail2.rol.it
Mime-Version: 1.0
Content-Type: multipart/related;
type="multipart/alternative";
boundary="=_12409671==_.REL"
NNTP-Posting-Date: 31 Jan 2002 11:56:55 GMT
Xref: pitt.edu sci.stat.edu:32797

--=_12409671==_.REL
Content-Type: multipart/alternative;
boundary="=_12409687==_.ALT"

--=_12409687==_.ALT
Content-Type: text/plain; charset="iso-8859-1"; format=flowed
Content-Transfer-Encoding: quoted-printable

Verba Volant 31-01-02,
Every day a new quotation translated into many languages.

<http://www.logos.net/bimbi/index_en.html>childrens_dictionary.png
_
Quotation of the day:
Author -=20
<http://www.google.it/search?q=3DElisa+F%E9lix&hl=3Dit&btnG=3DCerca+con+Goog=
le&lr=3D>Elisa=20
<http://www.google.it/search?q=3DElisa+F%E9lix&hl=3Dit&btnG=3DCerca+con+Goog=
le&lr=3D>F=E9lix

English - only that which I have lost belongs to me for always
Italian - soltanto ci=F2 che ho perso mi appartiene per sempre
Spanish - tan s=F3lo lo que he perdido me pertenece para siempre
French - seul ce que j'ai perdu m'appartient =E0 jamais
German - nur dass, was ich verloren habe, geh=F6rt mir ewig
Albanian - vet=EBm =E7'kam humbur m=EB p=EBrket p=EBrgjithnj=EB
Basque - galdu dudana bakarrik dut betiko nire
Bolognese - s=E5ul qu=E0ll ch=92ai =F2 p=EArs al m apart=E9n par
s=E4nper
Bresciano - s=F9l ch=E8l che go pird=ECt l=92=E8 l=92m=E9 per s=E8mper
Calabrese - sulu chiru ca 'gi=F9 persu =E8 miu pi sempri
Catalan - nom=E9s all=F2 que he perdut, ser=E0 meu per sempre
Croatian - samo ono =9Ato sam izgubila je moje zauvijek
Czech - jen to, o co jsem pri=9Ala, mi patr=ED nav=9Edy
Danish - kun det jeg har misted tilh=F8rer mig
Dutch - alleen wat ik heb verloren, behoort me voor altijd toe
Emiliano Romagnolo - sulment quel c'a io smarid l'=E8 par semper
cumpagn ma=
 me
Esperanto - nur tio kion mi estas perdinta apartenas al mi ciam
Ferrarese - sol qu=E9l ca i'o pers m'aparti=E9n par sempar
Finnish - vain se mink=E4 olen menett=E4nyt on ikuisesti minun
Flemish - alleen wat ik heb verloren, behoort me voor altijd toe
Galician - s=F3 aquilo que perd=EDn me pertence para sempre
Hungarian - csak az tartozik hozz=E1m mind=F6r=F6kre, amit m=E1r=
 elvesz=EDtettem
Latin - tantum quod perdidi meum in perpetuum est
Latvian; Lettish - tikai tas, ko esmu zaudejusi, man pieder uz
mu=9Eiem
Leonese - nam=E1i aqueillu que perd=ED pertenezme pa siempres
Mantuan - sol quel ch'=F2 pers 'l =E8 mio par s=E9npar
Mapunzugun - =F1i fij elk=FCnuel m=FCten inchegefuyem
Neapolitan - sulamento chello ca aggiu perduto =E8 pe ssempe d'o mio
Occitan - masque =E7=F2 qu=92ai perdut es miu per sempre
Parmigiano - soltant col che g'ho pers l=B4=E9 m=E9 per semper
Piemontese - mach l=F2n ch'i l'hai perd=F9 a l'=E9 m=E8 p=EBr sempe
Polish - tylko to co stracilem na zawsze mi przynalezy
Portuguese - s=F2mente isso que perdi sempre pertenecer=E1 a mim
Reggiano - sol c=F2ll ch'j=F2 peers l'=E9 mio per s=E8imper
Roman - solo quello che ho smarito =E8 mmio pe' sempre
Romanian - doar ceea ce am pierdut, =EEmi apartine pentru totdeauna
Sardinian (Limba Sarda Unificada) - petzi su chi apo p=E8rdidu m'at
a=20
apartenner pro semper
Sicilian - sulu chiddu c' haiu pirdutu...m'apparteni 'ppi sempri
Slovak - iba to, co som stratil, mi patr=ED nav=9Edy
Swedish - bara det jag f=F6rlorat =E4r mitt f=F6r all framtid
Turkish - sadece kaybettigim, her zaman bana aittir
Venetian - solo quel che go perso el sar=E0 mio par senpre
Zeneize - s=F6o che quello ch'=F2 perso o l'=E9 o m=E6 pe de longo
_

All languages, please click on this link
http://www.logos.net/owa-l/press.frasiproc.carica?code=3D508
_

To unsubscribe from Verba Volant, please follow this link:
http://www.logos.net/owa-l/press.rol_ml.verbavolant1?lang=3Den
and write in the empty field next to unsubscribe the email address
that you=
=20
find after "TO:" in the Verba Volant emails

alternatively write to the following address:
[EMAIL PROTECTED]=
=20
always co

Re: area under the curve

2002-01-30 Thread Rich Ulrich

On 30 Jan 2002 08:02:51 -0800, [EMAIL PROTECTED] (Melady Preece)
wrote:

 > A student wants to know how one can calculate the area under the
>curve for skewed distributions.  Can someone give me an answer about
>when a distribution is too skewed to use the z table?

It would be convenient if there was a general answer.

Here are some reasons why there isn't one answer:

 - Sometimes the accuracy of F(z)-for-F(x)  is needed within 
a certain absolute amount (like the accuracy for a KS 
goodness of fit).  Skewness does not index this.
 - Often, the accuracy of F(z)  is desired within 'so-many
decimals'  of accuracy:  two or three or more places of accuracy
for both  F(z) and 1-F(z).  Skewness does not index this.

In short: there are several ways to ask for accuracy, 
but there is no automatic (or pre-computed) connection that
I have heard of, between skewness and the accuracy of F, 
for any measure of accuracy.

I think I would find it interesting to see some Monte Carlo
experiments.  I think those should start with specific classes 
of distributions, and specific ranges for their parameters, and
plot their 'accuracies'  against the observed skewness 
and kurtosis.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: cutting tails of samples

2002-01-29 Thread Rich Ulrich

On 17 Jan 2002 00:05:02 -0800, [EMAIL PROTECTED] (HÃ¥kon) wrote:

> I have noticed a practice among some people dealing with enterprise
> data to cut the left and right tails off their samples (including
> census data) in both dependent and independent variables. The reason
> is that outliers tend to be extreme. The effects can be stunning. How
> is this practice to be understood statistically - as some form of
> truncation? References that deal formally with such a practice?

This is called "trimming" - 5% trimming, 25% trimming.
The median is what is left when you have done "50% trimming."

Trimming by 5% or 10% reportedly works well for your 
measures of 'central tendency', so long as you *know*  
that the extremes are not important.

I don't know what it is that you refer to as 'enterprise data.'

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: factor Analysis

2002-01-29 Thread Rich Ulrich

On Tue, 29 Jan 2002 10:52:30 +0100, "Huxley" <[EMAIL PROTECTED]> wrote:

> 
> Uzytkownik "Gottfried Helms" <[EMAIL PROTECTED]> napisal w wiadomosci
> [EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
> > It's not so simple. You have to do matrix-inversion for
> > that.
> >
> Not simple? I heard that taking suitable factor loadings and every variable
> mean I can obtain this space. e.g. (I do not know is it true)
> Let mean for car1 and questions 10 (variables):
> mean X1=1
> mean X2=2
> ..
> mean X10=10
> I have 2 factor score.
> factor loadins (aij) I have, therefore for first factor score, co-odrinate
> for car1 is
> F1(for car1)=1*a(1,1)+2*a(2,1)+3*a(3,1)+...+10*a(10,1)
> is it true?

No, that is not true.  
Please believe them.

Factor loadings are *correlations*  and serve as descriptors.  
They were neither scaled nor computed as regression coefficients -
which is what you are trying to use them as.


Now, in clinical research, we don't usually bother to create the
actual, real, true factor, for our practical purposes.   For practical
purposes, it is important to have some face-validity for what 
the factor means.  And it is handy for replication, as well as 
for understanding, if we construct a factor as the summed score
(or average score) of a set of the items.

So I look at the high loadings.  For a good set of items, it
can be realistic and appropriate to 'assign'  each item to the
factor where its loading is greatest, thus using each item just
once in the overall set of several derived factors.  (For a set 
of items where many items were new and untested, it can 
be appropriate to discard some of items -- where the loadings
were split, or were always small.)  Each factor is scored as 
the average score for of a subset of items.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Interpreting mutliple regression Beta is only way?

2002-01-25 Thread Rich Ulrich

On 18 Jan 2002 16:55:11 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> Rich Ulrich <[EMAIL PROTECTED]> wrote in message 
> 
> Thanks Rich, most informative, I am trying to determine a method of
> comparing apples to oranges - it seems an improtant thing to try to
> do, perhaps it is impossible .

Well, I do think there is a narrow, numerical answer about two
particular tests, in a particular sample.  But there are a lot
of side-issues before you get to an answer that is very useful.
See below.

> 
> I am trying to
> determine which is better, glycemic index or carbohydrate total in
> predicting glycemic load (Glycemic load=glycemic index*carbohydrate).
> 
> my results as a matrix:
> 
> GI load  GI  Carb
> GI load  1.000
> GI   .5331.000
> Carb .858.1241.000
> 
> So it seems that carb affects GI load more than does GI.. but this is
> on ALL foods.. (nobody eats ALL foods so cannot extrapolate to human
> diet) but I don't think you're allowed to do this kind of comparison
> as Carb and GI aretotal different values:
> 
> I suspected that you would be allowed to make the comparisons if you
> use Betas, ie. measure how many standard deviation
> changes of GI and  Carb it requires..  If it takes a bigger standard
> deviation of Carb then you could say that it is more likely that carb
> has a bigger effect on glycemic load.
> 
> you seem to suggest that even using standard deviation changes, you
> cannot compare  apples to oranges.  Which sounds right but is
> dissapointing..

There is a narrow, numerical thing with an answer.  For instance,
if you are adding A+B=C,  then two *independent*  components 
of C  "affect"   C  in proportion to their variances.  Your two 
components don't have much correlation (.13) -- that is, 
they are nearly independent -- so that would work out.

But you actually have an *exact* relationship, as a product.
The one that "matters"  in this case is the one that has the 
higher  "coefficient of variation" -- the larger standard 
deviation (and variance)  when expressed as logs.  
In a sum of two numbers, the one that is "more varying" 
will  "contribute more."

You can state that A  or B  is bigger, for *this* 
particular sample.  Now, I imagine a sample could be 
  a) all healthy; 
  b) all with a 'definite' diagnosis ; 
  c) all in the category of needing a diagnosis.
Or you could have some mixture of the above; for 
some specified ages, etc.  

Now, What is it you are trying to decide? - that will 
help determine a relevant sampling. 

When you say that a number is "more important", are 
you trying to say that the measure you have says
a lot?  - or that *improving*  that particular number
would gain you more, because the number you have
is a lousy one?  -- I can point out that if the two variables
matter exactly the same in some physiological sense, 
then you  can have two opposite conclusions about
which is "important," if one of them is *measured*  much
more poorly (much more inherent error; measurement 
error)  than the other.

And, you are reporting on tests, and a number.
You don't get to conclude that "glucose matters more 
than carbohydrates"  if you only know that your 50 mg
glucose test is more pertinent than your 50mg c.  test;
as collected in your  particular sample.


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: coefficient in logistic regression

2002-01-25 Thread Rich Ulrich

On 25 Jan 2002 08:13:41 -0800, [EMAIL PROTECTED] (Claudiu D. Tufis)
wrote:

> Hi, 
>  
> I have a multiple logistic regression. Among the predictors, I have 6
> variables that represent the dummies for an interaction term (the
> seventh is the reference category and is not included in analysis). I
> have obtained for five of these variables extremely large coefficients:
> exp(b) ranges from 90,000 to 166,000. 
>  
> Could you please tell me if it is normal to have such values for exp(b)?
> Do you think it is something wrong? 
>  

No, it is not normal.

Yes, something is wrong, if that is really what you have.

If you have more question, copy some output for us;
send some set-up lines; and mention what program.

Also, you posted many, many lines of HTML; 
please turn off that option if you can figure out how.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: QUERY on multiple linear regression: predicted values show much smaller ranges

2002-01-24 Thread Rich Ulrich

On 24 Jan 2002 07:09:23 -0800, [EMAIL PROTECTED] (Rich Einsporn)
wrote:

> Jim Clark gave a fine answer to the question posed by Sangdon Lee.
> However, I am curious about the correlation and R-square figures given by
> Sangdon.  Apparently, the R-squares for the simple linear regressions on
> X1 and X2 are (-.2)^2 = .04 and (.3)^2 = .09, but Sangdon says that the
> R-sq for the multiple regression is "ONLY" 0.3.  I find this to be
> surprisingly high, not low.  In the examples I see, the R-sq for the
> combined model is at most the sum of the individual R-squares. Is it even
> possible for the opposite to occur?

"Is it possible?"  Certainly.  The predictors have to be correlated 
for that to happen; and these were, at 0.6.  (Plug all the r's  into
the equation for multiple-R, and you can check his total.  I did not
check because it looked feasible, to my eyeball.)

"Confounding".  It is more common for two predictors to be 
highly correlated, and share their prediction variance, so 
that the total R^2  is barely more than either one alone.  
But these two variables, correlated 0.6 (which is pretty high),
predict in opposite directions; so their joint prediction will be
greater than the sum.

(This relates to another recent question:  At the extreme, 
this negative confounding is what can produce
standardized betas that are greater than 1.0.   My main
use of standardized betas is to compare them to the zero-
level r's  in order to judge confounding, in either direction.
I don't know of a regression statistic that directly tells me that;
maybe I should create one.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: analyzing counts from ordered categories

2002-01-23 Thread Rich Ulrich

On 23 Jan 2002 07:52:49 -0800, [EMAIL PROTECTED] (Mike Granaas) wrote:
[ snip ]
> 
>  ...  Their
> goal was then to test whether the pattern of endorsements was indeed
> 
> # item 1 > # item 2 > # item 3 > # item 4 > # item 5
> 
> where "# item 1" is short for "number of individuals endorsing item 1",
> and so on.
> 
> I know how to test for equality of proportions.  I even know how to test
> for specific hypothesized proportions.  But in this case there were no
> specified proportions, only a hypothesized ordering of proportions, I have
> no idea what to do.  
> 
> I could make up a set of proportions that fit with the hypothesis, but
> there are many such sets and so this seems hardly satisfactory.
> 
> Any ideas?
> 
> I am also leary of counting folks multiple times.  In this case if someone
> endorsed items 1, 2, and 3 they were counted in all three categories
> rather than, say, the highest category endorsed.  I have a vague
> recollection of some means of dealing with repeated measures count data,
> but certainly not recall what it was or under what circumstances it might
> be appropriate.

For endorsements of  #1 and #2, you have 
 -  people who endorsed neither; 
 -  people who endorsed both;  and
 -  people who endorsed one or the other.

If 1 & 2   are not ordered, 1-not-2  will equal 2-not-1.
So you can do McNemar's test on the two categories, 
to show whether one was endorsed more than the other.

Et cetera.  (There is also a multiple category test, but 
I am not sure whether it would fit.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: how to adjust for variables

2002-01-22 Thread Rich Ulrich

On 21 Jan 2002 16:53:31 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> Pretend you want to see how fat relates to cancer risk
> 
> fat   Kcalcancer
> 1 2   100
> 2 4   120
> 3 6   130
> 4 8   140
> 5 10  150
> 6 12  160
> 7 14  170
> 8 16  180
> 9 18  190
> 1020  200
> 
> You have to adjust for KCal, but how is this done, is the following
> the BEST way?

The problem is *nonsense*  as it is stated, 
since fat= KCal   except for the measurement units.

Adjusting A for A',  where A'  is approximately A 
except for irrelevant  measurement error, 
you have essentially nothing left.  Ever.

You always have to be careful when you adjust for 
something that has much correlation, to be sure that
the direction of the logic thoroughly makes sense.
(If there is not much correlation, then there is not 
much change possible in the estimator -- though
a *test*  could become more significant if a bunch
of error variance is accounted for.)

A simple way to "adjust for sex" (for instance), is
to compute  a statistic for each sex separately, 
and then average.  "Within group" is the basic 
idea of adjusting.

[ ... ]
> Is doing a univariate regression between the variable you want to
> adjust for and your predictor the only way to adjust for values as

Univariate?  Absolutely not.  *Multiple*  regression gives 
"partial regression coefficients."   Those  "adjust."

> above?  Studies often cite how they have "adjusted" for KCal, is this
> the way they do it, they usually do not specify the method.

The method is there, ordinarily, in the articles I read.
It is a poor journal (I think..., I estimate blindly ...) that
does not require a statement of that method, but if you
know nothing of regression, etc., you won't recognize
the statement when it appears.
 
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: 95% CI for a sum

2002-01-22 Thread Rich Ulrich

On 21 Jan 2002 08:55:21 -0800, [EMAIL PROTECTED] (Scheltema,
Karen) wrote:

> I have 2 independent samples and the standard errors and n's associated with
> each of them.  If a and b are constants, what is the formula for the 95%
> confidence interval for 
> (a(Xbar1)+b(xbar2))?
> 

Var(A+B) = Var(A) + Var(B)for A,B independent;

and
Var(kA) = k^2  Var(A)

Right?  The SE  is the SD of the mean, 
so you have the terms you need.

-- 
Rich Ulrich, [EMAIL PROTECTED]

http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Likelihood ratios and non-nested models

2002-01-20 Thread Rich Ulrich

On 19 Jan 2002 10:25:10 -0800, [EMAIL PROTECTED] (Brian Leung)
wrote:

> Greetings,
> 
> I've read that the likelihood ratio test is not valid for non-nested
> models. Is this still true if the PF (i.e., multinomial) is the same,
> but the link function differs.
> 

Yes, it is still true.  It is the subtraction that allows the 
estimates to be  *independent*  so that the result may be 
readily interpreted.

However, it is also true that people decide and publish using
related criteria for  non-nested models.  Such conclusions 
are not as rigorous, but seem to work in various applications.
Your comparison of link functions sounds familiar, for instance.

Here is one link I found by searching on 
< BIC "information criterion" > ,
  http://www.saam.com/faq/saam2/right.html
(I keep forgetting how to spell AKAIKE.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Interpreting mutliple regression Beta is only way?

2002-01-16 Thread Rich Ulrich

On 16 Jan 2002 11:33:15 -0800, [EMAIL PROTECTED] (Wuzzy) wrote:

> If your beta coefficients are on different scales: like
> you want to know whether temperature or pressure are affecting
> your bread baking more,
> 
> Is the way to do this using Beta coefficients calculated
> as Beta=beta*SDx/SDy

Something like that ... is called the standardized beta, 
and every OLS  regression program gives them.
 ...
> 
> It seems like the Beta coefficients are rarely cited in studies
> and it seems to me worthless to know beta (small "b") as you are
> not allowed to compare them as they are on different scales.

In biostatistical studies, either version of beta is pretty worthless.
Generally speaking.

What you have is prediction that is barely better than chance.
The p-values tell you which is "more powerful"  within this one
equation.  The zero-level correlation tells you how they 
related, alone.  -- If these two indicators are not similar, then
you have something complicated going on, with confounding
taking place, or joint-prediction, and no single number will show
it all.
 - When prediction is enough better-than-chance to be 
really interesting, then the raw units are probably interesting, too.

 [ ... ]
> Is there a way of converting this standardized coefficient to a
> "correlation coefficient" on a scale of -1 to +1)
> It would be useful to do this as you want to know the correlation
> coefficient of temperature after factoring out pressure.

I think you are looking for simple answers that can't exist, 
even though there *is*  a partial-r, and the beta in regression
*is*  a partial-beta.

The main use I have found for the standardized (partial) beta
is the simple check against confounding, etc.  If ' beta'  is similar
to the zero-order r,  for all variables, then there must be pretty
good independence among the predictors, and interpretation
doesn't hide any big surprises.  If it is half-size, I look for shared
prediction.  If it is in the wrong direction or far too big (these 
conditions happen at the same time, for pairs of variables), 
then  gross confounding exists.

Hope this helps.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: a problem

2002-01-16 Thread Rich Ulrich

On Tue, 15 Jan 2002 23:06:25 GMT, janne <[EMAIL PROTECTED]> wrote:

> Lets say I do a x2(chi) test and have the hypothesis:
> 
[snip, some example]
> 
> If you can have < in hypothesis, then when is it < and when is it > I
> should use? How do I know which one to use?
> 
> I also wonder about t-tests the same question. When do I know if I
> should use < or >?

Are you referring to the chisquared test on a contingency table?  
That is the most popular thing called  'chi-squared test'  but it is
far from the only thing.  

Almost always (but not always), chisquared is used to 'reject'
when the chisquared value is large.  Now, if your H0  and Ha
require "less than,"  that is a pretty good indicator that you 
should *not*  be using the contingency table; but you might
be using a 'test for proportions'  that gives you  the square
root of a 1-d.f.  chi-squared, which is a normal-deviate z:  which
has either a plus or minus sign attached.  But you *can* use
the chisquared test, and make sure the differences are in
the right direction.

Short answer:  Look at the numbers, and use your head.
There is not a magical formula that makes a stupid-looking
answer come out to be correct.

If this still seems confusing, borrow a book or two on 
experimental design and spend time on the earliest chapters.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Empirical data Fitting

2002-01-15 Thread Rich Ulrich

[ rearranging to the usual order, with Reply at the bottom ]

> Chia C Chong wrote:
> > 
> > Hi!!
> > 
> > I have a set of data with some kind of distribution. When I plotted the
> > histogram density of this set of data, it looks sth like the
> > Weibull/Exp/Gamma distribution. I find the parameters that best fit the data
> > and then, plot the respective distribution using the estimated parameters on
> > the empirical distribution. My question is, what kind of statistical test
> > that I should  ...
   [ ... ]

On 14 Jan 2002 08:15:35 -0800, [EMAIL PROTECTED] (kjetil
halvorsen) wrote:

> A quantile-Quantile plot for graphical comparison is best, if you need a
> numerical test you can use the pearson correlation coefficient between
> the observed and expected quantiles. A table for that test you can  ake
> for yourself with simulation. 

No, I think that advice is not good, in this instance.  Someone 
else has mentioned that these three belong to the same family 
so that it is possible and most desirable to solve for the value 
of the parameter that distinguishes them.

The correlation between those 'quantiles' -- That (I think)
gets you a test like the Shapiro-Wilks  which is a fine test for 
normality when you compare correlations across a range of
deviant samples.  I don't remember its analog being used 
for testing between theoretical  CDFs; and I suspect it is inferior
to other approaches for (say) strongly skewed distributions.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Fisher & Tippett Extreme Value Theorem

2002-01-13 Thread Rich Ulrich

On 10 Jan 2002 03:28:29 GMT, [EMAIL PROTECTED] (Vmcw) wrote:

> Hi!
> 
> I'm looking at an article that references a result by Fisher & Tippett
> (specific reference is "Limiting Forms of the Frequency Distribution of the
> Largest or Smallest Member of a Sample," Procedings of the Cambridge
> Philosophical Society 24, 180-190 (1928)).  The article presents the result as
> follows:  if there are S independent samples of size m from a parent population
> bounded below by a, and if xi is the smallest value in sample i, then the
> limiting distribution of the minimum of the xi-values is Weibull with location
> parameter a as m -> infinity.
> 
> This seems like a not particularly rigorous description of the result, and I
> wanted to read the original statement to make sure that all the details are
> covered here, i.e. it seems like the parent distribution would have to be
> continuous, at least, and that a should be the greatest lower bound rather than
> just a bound.  So, I ordered the original article from my library and have been

The way that I read your 1928  summary, I don't figure out why you 
are looking for a 'bound.'   It says, some distribution becomes 
(increasingly) Weibull.  With a parameter.  Where do you see a bound?

And, yes, I would guess that somewhere they say 
something is continuous.

> waiting a lng time with no success.  So... can anyone help me out with a
> precise statement regarding this result?  (If it helps, I have obtained
> Gumbel's "Statistics of Extremes" to see if it's buried in there, but so far
> haven't found it.  Page number?)
> 
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: SAT Question Selection

2002-01-13 Thread Rich Ulrich

On Sun, 13 Jan 2002 13:04:14 GMT, "L.C." <[EMAIL PROTECTED]>
wrote:

> Back in my day (did we have days back then?) I recall
> talk of test questions on the SAT. That is, these questions
> were not counted; they were being tested for (I presume)
> some sort of statistical validity.
> 
> Does anyone have any statistical insight into the SAT question
> selection process. Does anyone have a specific lead? I can
> find virtually nothing.

I believe that they have to change their questions a lot more
often than they used to, now that they occasionally reveal 
some questions and answers.

The Educational Testing Service has a web site that looks
pretty nice, in my 60-second opinion.
   http://www.ets.org/research/

They do seem to invite communication -- I suggest you e-mail,
if you don't find  what you are looking for in their 8 research areas,
or elsewhere.

It seems to me that I found a statistics journal produced by ETS  
when I was looking up references for scaling, a year or so ago.
But I don't remember that for a fact.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Buy Book on "Probability and statistical inference"

2002-01-13 Thread Rich Ulrich

On Sat, 12 Jan 2002 14:37:10 -, "Chia C Chong"
<[EMAIL PROTECTED]> wrote:

> Hi!
> 
> I wish to get a book in Probability and statistical inference . I wish to
> get some advices first..Any good suggestion??

 a) Browse in a good college bookstore.  There should be a 
section among general books, in addition to the textbooks 
that are filed by course.  You can not the titles, even if the
books are sealed in plastic.  Anything that has been in print
for decades is good (for something).
 b) Look for some of those titles in a good college library, 
and go to the shelves where you would find them.  Browse
the contents to see what you are looking for.
 - Then you could ask a stat-group for informed opinions, 
if you don't get titles in response to this first request.

Further, online:  you could do a http://groups.google.com  search 
of the stat-groups.  You could go to stat-web pages of people who 
answer questions in the stat-groups, and look for references and 
links  to references.

Hope this helps.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Proportionate vs. disproportionate

2002-01-11 Thread Rich Ulrich

On 11 Jan 2002 07:46:20 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

> definition: proportionate = equal % change
> 
> IF we agree on this ... and maybe we don't ... then, since the % change is 
> always UN =, then all changes are DISproportionate
[ ... ]

Are you sure you *advocate*  that?  

I am not positive, but
I think I would have objected to "equal % change"  
as  =proportionate=  by the time I finished algebra in high school. 

I know I have objected to similar confusion, on principled 
grounds, since I learned about  Odds Ratios.

I suspect that the original sample was small enough that 
the apparent difference in ORs   was not impressive.
-- 
RIch Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: what is the appropriate method?

2002-01-10 Thread Rich Ulrich

On Thu, 10 Jan 2002 12:00:06 +0100, "Jos Jansen"
<[EMAIL PROTECTED]> wrote:

> 
> "Rich Ulrich" <[EMAIL PROTECTED]> wrote in message
> [EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
> > On Wed, 09 Jan 2002 08:33:41 GMT, [EMAIL PROTECTED]
> > (Jukka Sinisalo) wrote:
> >
> > >
> > > We have two pots with 25 plants each. After an identical treatment we
> > > wait for a week, and then calculate how many of the plants died. We
> > > repeat this "experiment" 20 times, so we end up with 20 pairs of
> > > "survival percantages".
> > >
> > > We are interested in determining the accuracy/reliability of our
> > > method.  In other words if in the future we use just one pot of 25
> > > plants, what will be the confidence interval of the result.
> 
> snip
> 
> > "transform the data"  - is easy and apt.
> >
> > Compute the logit and use that in your modeling.  With
> > the difference of two of them, you have the "log Odds Ratio."
> 
> snip
> 
> Using logits is obvious, but log Odds Ratio is not, given the aim to use
> only one pot in the future (not the difference of two). An estimate of the
> sum of variance components within and between repeats will be required for
> calculating the precision of a single result.
> 

Oh, right.  Thanks.  
I was reading the problem as something different.  Misreading it.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: what is the appropriate method?

2002-01-09 Thread Rich Ulrich

On Wed, 09 Jan 2002 08:33:41 GMT, [EMAIL PROTECTED]
(Jukka Sinisalo) wrote:

> 
> We have two pots with 25 plants each. After an identical treatment we
> wait for a week, and then calculate how many of the plants died. We
> repeat this "experiment" 20 times, so we end up with 20 pairs of
> "survival percantages".
> 
> We are interested in determining the accuracy/reliability of our
> method.  In other words if in the future we use just one pot of 25
> plants, what will be the confidence interval of the result.
> 
> At the moment we calculate the  pairwise differences,
> and use the standard deviation of those 20 differences to
> estimate the uncertainty of our method.  The standard deviation
> I end up is relatively large, and if the survival percentage happens
> to be high, I get a funny confidence interval where the upper limit
> is > 100%, which is clearly impossible.
> 
> I have a feeling this is not the best way to go about this,
> but I'm unsure what to do. Perhaps use maximum likelihood
> probabilities for P("plant dies"), or somehow transform the
> data?

"transform the data"  - is easy and apt.

Compute the logit and use that in your modeling.  With
the difference of two of them, you have the "log Odds Ratio."

The logit is a natural for modeling growth-within-limits;
and it works rather well for modeling percentages.
For your example, you will probably end up talking 
about the log of the Odds Ratio  for each  week.

Among other advantages, it has this one that you mention:
When you back-transform, the  CIs  are never improper.  

The logit  is  < log( p/(1-p) ) >
You can't compute LOG  if the survivorship is ever 0 or 100%.  
Folks usually plug in  0.5  for 0 cases when that happens.
(Starting with 25 plants,  plug in 2% or 98% for 0 or 100%.)


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Chart of the Week

2002-01-09 Thread Rich Ulrich

On Wed, 09 Jan 2002 00:47:19 -0500, Derrick Coetzee
<[EMAIL PROTECTED]> wrote:

> Gary Klass wrote:
> 
> > http://lilt.ilstu.edu/gmklass/COW/
> 
> As other posters have noted: always beware the "obvious" implications of 
> correlations. A common example is that drownings and ice cream sales are 
> strongly correlated.
 [...]

Yesterday's newspaper summarized a report that screen-writers 
who had more movies to their credit, and more 4-star pictures 
in particular, were more likely to have won a writing Oscar, 
instead of just being nominated.   No! Wait!  It was backwards 
of that -- the authors (so it seemed in the news item)  concluded 
that Winners should go on to be more successful...  but die younger.

And further speculation.  My interest faded before I looked it all up.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Is there a problem with this kind of comparison?

2002-01-04 Thread Rich Ulrich

On 3 Jan 2002 20:17:48 -0800, [EMAIL PROTECTED] (Lucas Wells)
wrote:

[ snip, detail ]

> 
> Now, I look at these percentages and I think to myself, 'They're
> percentages of a whole. If one goes up, then another must fall. It
> doesn't seem to make sense to examine them as if they are measures
> that can be seperately influenced (ie, as if we could decrease
> percentages across the board).'
> 
> Is this a legitimate concern?
> 
> I could understand it, if one type of error was more 'important' than
> another, then perhaps you would be trying to minimise the percentage
> of that particular error, but you would expect the others to inflate
> as a result, yes?


Absolutely right.  For the numbers given, the report is silly.  
(Your example has just one error per-thing.  Does the big 
report look much different with multiple errors?)

There could be rates of errors per transaction;  per
section-of-a-transaction; per total-errors.  Whatever exists.

The rate or fraction of errors per transaction  
seems to be a standard that would be worth tracking across
reports.  

The fraction per all-ERRORs   seems pretty useless, unless,
for instance, you were in a bad state where (say)  90% of all 
errors were one kind, so you were trying to bring that one 
sort down.

Hope this helps.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Empirical data Fitting

2002-01-03 Thread Rich Ulrich

On Thu, 3 Jan 2002 01:02:17 -, "Chia C Chong"
<[EMAIL PROTECTED]> wrote:

> Hi Bill..Thanks for your reply. You mentioned in the last line of your
> message that statistical tests are not a very good way to choose among
> distributions. If this is the case, what test do you think is better in my
> case??
 ...

Other that statistical?  
 - Being familiar;  being logically inherent;  being brief...   

Occam's razor goes something like this,
"Do not unnecessarily multiply entities in an explanation."

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: One-tailed, two-tailed

2002-01-02 Thread Rich Ulrich


On Sun, 30 Dec 2001 18:07:16 -0500, [EMAIL PROTECTED] (Stan Brown)
wrote:

> Rich Ulrich <[EMAIL PROTECTED]> wrote in sci.stat.edu:
[ ... ]
RU > >  We should
> >not overlook the chance to teach our budding statisticians:
> >*Always*  pay attention to the distinction between random trials 
> >or careful controls, on the one hand; and grab-samples on the other.
> >[Maybe your teacher asked the question that way, in order to
> >lead up to that in class?]
> 
SB > [ snip; was in a book of homework problems ...]
RU > 
> >The numbers do  not  *prove*  that one gas gives better mileage;
> >the mileage was, indeed, better for one gas than another -- for
> >reasons yet to be discussed.  Different cars?  drivers?  routes?

SB > 
> All good points for discussion. But I wouldn't focus too much on 
> that off-the-cuff word "prove". (I'm not being defensive since I 
> didn't write the exercise. :-) My students did understand that 
> nothing is ever proved; that there's still a p-value chance of 
> getting the sample results you got even if you did perfect random 
> selection an d the null hypothesis is true. Maybe I'm being UNDER-
> scrupulous here, but I think it a pardonable bit of sloppy language.
> 

I read this, and I re-read it, and it still seems like Stan has 
missed the gist.  Type  I  error ?   Stan sees all the problem
(so it seems to me)  as being one of Type I error, "there's 
still a p-value chance..."
Yes, you  *might*   construe *some* of the problem that way, 
some of the time.

But I am  concerned with the fundamental difference,
randomized trials  versus  grab-sample (observational):
For the former, you get to start with the assumption that 
outside variables are not important.  For the latter, you
*have to*  start with the assumption -- if you are a properly
trained scientist -- that they are *apt*  to be.

It is not enough that you have a 'single, important hypothesis'
like you have in a design;  you have to explain away the
other possible causes.  

RA Fisher (along with a few others) defended tobacco
for years.  The case against tobacco was not solidly made
until *other*,  competing explanations had been laid to rest.
That included a convergence with biomedical evidence.


I was annoyed by  sloppiness when I was in grad-school.
If I remember right, it was the epidemiologists who 
regularly were sloppy, and failed to qualify their tests,
while the math-trained statisticians were apt to be clear.
"We conclude that these *numbers*  are different..."
instead of "... conclude that these *samples*  are different."

 - I mention this because it suggests to me  that such
sloppiness is *not*  pardonable;  that the people who care
most about the 'actual outcomes'  are the ones who are 
least able to keep in mind the technical reservations.

That's my advice, for what it is worth -- I have never taught
any regular course.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: One-tailed, two-tailed

2001-12-30 Thread Rich Ulrich

[ posted and e-mailed.]
On Sat, 29 Dec 2001 16:46:10 -0500, [EMAIL PROTECTED] (Stan Brown)
wrote:

[... ]
> 
> [The student is led to enter two sets of unpaired figures into 
> Excel. They represent miles per gallon with gasoline A and gasoline 
> B. I won't give the actual figures, but here's a summary:
> 
>   A: mean = 21.9727, variance = 0.4722, n = 11
>   B: mean = 22.9571, variance = 0.2165, n = 14
> 
> The question is whether there is a difference in gasoline mileage. 
[snip, about variances.]
> 
> "What's your conclusion about the difference in gas mileage?" 
> [Answer: At significance level 5%, previously selected, there is a 
> difference between them.]
> 
> Now we come to the part I'm having conceptual trouble with: "Have 
> you proven that one gas gives better mileage than the other? If so, 
> which one is better?"
> 
> Now obviously if the two are different then one is better, and if 
> one is better it's probably B since B had the higher sample mean. 

I want to raise an eyebrow at this earlier statement.  We should
not overlook the chance to teach our budding statisticians:
*Always*  pay attention to the distinction between random trials 
or careful controls, on the one hand; and grab-samples on the other.
[Maybe your teacher asked the question that way, in order to
lead up to that in class?]

The numbers do  not  *prove*  that one gas gives better mileage;
the mileage was, indeed, better for one gas than another -- for
reasons yet to be discussed.  Different cars?  drivers?  routes?

> But are we in fact justified in jumping from a two-tailed test (=/=) 
> to a one-tailed result (>)?
> 
> Here we have a tiny p-value, and in fact a one-tailed test gives a 
> p-value of 0.0001443. But something seems a little smarmy about 
> first setting out to discover whether there is a difference -- just 
> a difference, unequal means -- then computing a two-tailed test and 
> deciding to announce a one-tailed result.

Another small issue.  Why did the  .00014  appear?  

In clinical trials, we observe the difference and then we do attribute
it to one end.  But it is not the convention to report the one-tailed
p-level, after the fact.  I think there are editors who would object
to that, but that is a guess.  Also, for various reasons, our smallest
p-level for reporting is usually  0.001.


> 
> Am I being over-scrupulous here? Am I not even asking the right 
> question? Thanks for any enlightenment.
> 
> (If you send me an e-mail copy of a public follow-up, please let me 
> know that it's a copy so I know to reply publicly.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Is this how you would have done it?

2001-12-27 Thread Rich Ulrich

On Sat, 22 Dec 2001 09:02:23 -0500, "Ralph Noble"
<[EMAIL PROTECTED]> wrote:

 ... 
> A local newspaper asked its readers to rank the year's Top 10 news stories
> by completing a ballot form. There were 10 choices on all but one ballot
> (i.e. local news, sports news, business news, etc.), and you had to rank
> those from 1 to 10 without duplicating any of your choices. One was their
> top pick, 10 their lowest. Only one ballot had more than 10 choices, because
> of the large number of local news stories you could choose from.
> 
> 
> 
> I would have thought if you only had 10 choices and had to rank from 1 to
> 10, then you'd count up all the stories that got the readers' Number One
> vote and which ever story got the most Number One votes would have been
> declared the winner.
 [ ... ]

I have read three good responses.

I want to mention that what you describe is just like the polls 
used in college sports, where the coaches or  media each vote
(coaches do one poll; media do another)   for 
"Who is number 1."And they typically do report what you
ask for, the number of  #1 votes, in addition to the total (which
does not have to agree).

Thinking of other ratings with rankings:

Is it "Places rated almanac"? - that annually lists numbers 
for 120 or 250  American cities.  They do another simple
averaging of ranks, across their 10  or so categories.

I remember some public discussion of how arbitrary 
that was, and how it tends to reward 'consistent mediocrity.'
(The first time they did this, Pittsburgh  was near 
the top, so I have noticed later discussion.)

That discussion also pointed out that the results 
were  *usually*   not  *greatly*   different if you 
chose another weighting system.  
And the better advice was that any
interested person should  look at the separate categories, 
and choose the few that matter to themselves.


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: How ro perform "Runs Test"??

2001-12-27 Thread Rich Ulrich

On Sun, 23 Dec 2001 23:48:58 GMT, "Jim Snow" <[EMAIL PROTECTED]>
wrote:

> 
> "Glen" <[EMAIL PROTECTED]> wrote in message
> [EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
> > [EMAIL PROTECTED] (Chia C Chong) wrote in message
> news:<[EMAIL PROTECTED]>...
> > > I am using nonlinear regression method to find the best parameters for
> > > my data. I came across a term called "runs test" from the Internet. It
> > > mentioned that this is to determines whether my data is differ
> > > significantly from the equation model I select for the nonlinear
> > > regression. Can someone please let me know how should I perform the
> > > run tests??
> >
> > You need to use a runs test that's adjusted for the dependence in the
> > residuals. The usual runs test in the texts won't apply.
> >
> > Glen
> 
> I always understood that the runs test was designed to detect systematic
> departures from the fitted line because some other curve fitted the data
> better. In this context, it is a test for dependence of residuals.
> 
> There is a discussion of this at
>  http://216.46.227.18/curvefit/systematic_deviation.htm
> 
> Any elementary text in Non-parametric Methods in statistics will
> give an example.

Well, the residuals are always *dependent*, to the extent of p/n  
(# variables  divided by N).  That is the Expectation.  So they are
*not*   i.i.d, which is an assumption.   Thus:  the runs test is an
approximation which is inadequate for large ratios of p/n -- It 
is nice for the stat-pack to explain the runs-test, but 
not-so-nice that it fails to mention the other detail.

Draper and Smith's book on regression mention that the runs
test will be approximate, since the expectation is not independent.

You can also google-search on  <"Durbin-Watson"  "runs test">,
and click on the lectures ...  or whatever appeals most to you.
The D-W  test is awkward enough to *test*  that you don't wonder 
why people should look for an easier option.  Several textbooks 
that I just looked at seem to be satisfied with recommending 
that you eye-ball your residuals in several plots - without doing
tests.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What is the formulae for generalised Gaussian CDF???

2001-12-21 Thread Rich Ulrich

On 15 Dec 2001 06:41:29 -0800, [EMAIL PROTECTED] (Chia C Chong) wrote:

> Can somebody tell mw what is the formulae for the CDF of generalised
> Gaussian distribution or where can I refer to???
> 


http://amp.ece.cmu.edu/Publication/Trista/ACMSecurity_Trista.pdf

 - has an equation.  That equation cites

[16] J. R. Hernández, M. Amado, and F. Pérez-González, "DCT-
Domain Watermarking Techniques for Still Images: Detector 
Performance Analysis and a New Structure". 
IEEE  Transactions on Image Processing, 9(1), 55-68. 
 

That was one of the first files I looked at after a 
www.google.com  search on "generalized gaussian"; 
so, you can probably find others.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Standardizing evaluation scores

2001-12-19 Thread Rich Ulrich

On Wed, 19 Dec 2001 02:11:01 GMT, Doug Federman <[EMAIL PROTECTED]>
wrote:

> I have a dilemma which I haven't found a good solution for.  I work with 
> students who rotate with different preceptors on a monthly basis.  A 
> student will have at least 12 evaluations over a year's time.  A 
> preceptor usually will evaluate several students over the same year.  
> Unfortunately, the preceptors rarely agree on the grades.  One preceptor 
> is biased towards the middle of the 1-9 likert scale and another may be 
> biased towards the upper end.  Rarely, does a given preceptor use the 1-9 
> range completely.  I suspect that a 6 from an "easy" grader is equivalent 
> to a 3 from a "tough" grader. 

Huge rater-differences?  that implies that you can't keep the
original 1-9  and be fair.  Doesn't it?  And you can't trust the
labels stuck with the original 1-9.

> 
> I have considered using ranks to give a better evaluation for a given 
> student, but I have a serious constraint.  At the end of each year, I 
> must submit to another body their evaluation on the original 1-9 scale, 
> which is lost when using ranks.
> 
> Any suggestions?

The original 1-9  seems to have no rational anchor, if the
raters can vary between by 3 points on a standard.  
So, you average your ranks; 
sort the students that way; 
compute and list their raw averages along with rank averages.

If extreme HIGH scores or LOW scores are important, 
you could avoid ranks.
(a) Simply rescore each preceptor to have an average of 5, or 
(b) normalize each preceptor by mean and standard deviation.

Assume ranking.
I think, from that step, I might assign the 
mid-average to the mid-rank -- this would have
the effect of anchoring my eventual, final report
to the same range the raters used.  (Assuming this
matters.)  Then I would work my way to the extremes, 
'grading by the curve.'   -- If you are trying to respect
the labels on the 1-9 Likert, you have to consider
what they actually *say*.  You might justify giving 
the best  student a '9'  despite an average of 6.9  if 
the label says "best, consistent performance."

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: chi square validity?

2001-12-18 Thread Rich Ulrich

On Tue, 18 Dec 2001 14:19:34 + (UTC), [EMAIL PROTECTED]
(Benjamin Kenward) wrote:

> 
> Let's say you have a repeatable experiment and each time the result can be
> classed into a number of discrete categories (in this real case, seven).
> If a treatment has no effect, it is known what the expected by chance
> distribution of results between these categories would be. I know that a
> good test to see if a distribution of results from a particular treatment
> is different to the expected by chance distribution is to use a
> chi-squared test. What I want to know is, is it valid to compare just one
> category? In other words, for both the obtained and expected
> distributions, summarise them to two categories, one of which is the
> category you are interested in, and the other containing all the other
> categories. If the chi-square result of the comparison of these categories
> is significant, can you say that your treatment produces significantly
> more results in particularly that category, or can you only think of the
> whole distribution?

Mathematically, the statistical test is okay:  There is no 
problem if you decided at the outset that 7 categories 
should be scored as D-and-not-D, so you would do a 
2x2  contingency table test.

Other inferences, of course, are more problematic.  "Multiple-tests."
Deciding on a test based on the outcomes is a form a cheating
in the hypothesis testing, if you don't take that into account in
the reporting of it.

If your overall test is significant -- with 6 d.f.,  I think -- then
it is somewhat conventional to look at the separate contributions 
by cell, without being too shy.  If the overall test is *not*  that 
happy, then you ought to state that, and offer further guesses
as purely exploratory or suggestive numbers.  Then you can 
describe one cell's contribution "versus the others."  

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Transaction Anomaly on 9/11????

2001-12-18 Thread Rich Ulrich

On Mon, 17 Dec 2001 15:20:55 GMT, [EMAIL PROTECTED] (J. Williams)
wrote:

> I recently read about German-based Convar helping cos. in NYC uncover
> the facts surrounding the unusual surge (both in volume and amounts)
> in financial transactions during & immediately preceding the WTC
> disaster.  Convar is using a laser scanning technology to recover data
 [ snip ]

Do you have a citation for your "unusual surge"? 

In my daily newspapers (NY Times, for one), I read
the opposite -- there was no indication at all that the 
plans of the terrorists had leaked out to speculators, 
of any ilk.  And investigators looked.

I'm sorry, but this post reads to me like a pandering
to hostile paranoia, and that annoys me;  so I won't 
say anything more.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Basics

2001-12-14 Thread Rich Ulrich

On Fri, 14 Dec 2001 09:22:26 +1100, "colsul" <[EMAIL PROTECTED]>
wrote:

> Read the request closely as answering newsgroup queries without
> understanding what is said can make you look, no, confirm you are stupid.

In Sensitivity-group country, we try to say (if confrontation is
necessary)  that an *action*  is stupid or poorly considered, not 
that a person is stupid.  (That is usually enough to make one's
point...  without getting embarrassed by over-ambitious flaming)

In sci.stat.edu, we usually post the content that we are replying
to, before we add comments; and we might criticize in some fashion,
but we would never call a regular post-er 'stupid',  who has offered 
dozens or hundreds of useful comments in the past.

Before posting to a particular Internet group, it is 
well-advised to read a number of posts there, to pick up on
the acceptable content and style.

Besides all that, I have to say: from what I made of your question,
I agree 100% with Glen's response.  He gave a (relatively) gentle
warning about a course of action.

As a matter of fact, I find it hard to *imagine*  cramming (which 
you ask about) for a job interview,  unless it was following the
advice,  'Learn about the job/ company.'   But that's not 
something to look up in a textbook.


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students: A Statistical Perspective

2001-12-13 Thread Rich Ulrich

On 7 Dec 2001 14:24:17 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

> At 08:08 PM 12/7/01 +, J. Williams wrote:
> >On 6 Dec 2001 11:34:20 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:
> >
> > >if anything, selectivity has decreased at some of these top schools due to
> > >the fact that given their extremely high tuition ...
> 
> 
> i was just saying that IF anything had happened ... that it might have gone 
> down ... i was certainly not saying that it had ...
> 
> but i do think that it could probably not get too much more selective ... 
> so it probably has sort of stayed where it has over the decades ... so if 
> grade inflation has occurred there it would not likely be due to an 
> increased smarter incoming class
> 
(In the NY Times)  At Harvard in particular, the interviewees
claimed that the present freshmen had notably better SATs
than those of a generation ago -- There are not nearly so 
many people with so-so scores (alumni offspring?), and a 
quarter of the class now has SATs that are perfect 1600, 
or nearly that (?no explanation of what 'nearly' means).

I go along with the notion that, in the long run,  if there is to 
be special meaning to being an "honors graduate" from Harvard,  
it can't mean "top 75% of the class".  (I think that is what 
someone reported, somewhere.)

I remember reading, years ago, that the Japanese school
trajectory differed from ours -- they learnt a lot before college,
and college was a long party before starting a career.  (This
was a few years ago.)  Their life-long success was pre-determined
largely by which-university accepted them; it sounded like 
the old-school-tie was a huge social asset.

Reportedly, that was why their high school students worked 
so hard on cram courses and extra studying; college was 4 
years of party.  - Since they are sliding away from lifetime
employment, etc., I wonder if the educational system is
becoming more flexible and technocratic, too.  Are  our systems 
converging?

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When to Use t and When to Use z Revisited

2001-12-10 Thread Rich Ulrich

On Mon, 10 Dec 2001 12:57:29 -0400, Gus Gassmann
<[EMAIL PROTECTED]> wrote:

> Art Kendall wrote:
> 
> (putting below the previous quotes for readability)
> 
> > Gus Gassmann wrote:
> >
> > > Dennis Roberts wrote:
> > >
> > > > this is pure speculation ... i have yet to hear of any convincing case
> > > > where the variance is known but, the mean is not
> > >
> > > What about that other application used so prominently in texts of
> > > business statistics, testing for a proportion?
> 
> > the sample mean of the dichotomous (one_zero, dummy) variable is known, It
> > is the proportion.
GG > 
> Sure. But when you test Ho: p = p0, you know (or pretend to  know) the
> population variance. So if the CLT applies, you should use a z-table, no?
> 

That is the textbook justification for chi-squared and z  tests
in the sets of 'nonparametric tests'  which are based on 
rank-order transformations and dichotomizing.

The variance is known, so the test statistic has the shorter tails.
(It works for ranks when you don't have ties.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What usually should be done with missing values when I am conducting a t test or other tests?

2001-12-07 Thread Rich Ulrich

On Fri, 07 Dec 2001 04:59:46 GMT, Richard J Burke
<[EMAIL PROTECTED]> wrote:
> jenny wrote:
> 
> > What should I do with the missing values in my data.  I ned to perform
> > a t test of two samples to test the mean difference between them.
> >
> > How should I handle them in S-Plus or SAS?
> >
> > Thanks.
> > JJ
> 
> If you are doing paired tests, then the pairs with missing values will
> have to be ignored; otherwise, you will simply have two samples of
> different sizes.

Going a step further --

The awkward dataset has *some* overlap from test-1 to test-2,
and most Subjects exist just in one set or the other.  But there
are enough pairs to test.

Then you have the paired test, and the unpaired test; and 
you can test assumptions about whether either set might
me missing non-randomly.  
(And if everything looks kosher, you can sum the two tests.)
But you should try to ask everyone involved, 
"Why and how do you happen to have data like these?"
 - If there is a Why, you need to find out Why; 
else, you want to educate them to do better next time.

-- 
Rich Ulrich, [EMAIL PROTECTED]


http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students: A Statistical Perspective

2001-12-06 Thread Rich Ulrich

Just in case someone is interested in the Harvard instance 
that I mentioned -- while you might get the article from a newsstand
or a friend --

On Sun, 02 Dec 2001 19:19:38 -0500, Rich Ulrich <[EMAIL PROTECTED]>
wrote:

[ ... ]
> 
> Now, in the NY Times, just a week or two ago.  The
> dean of undergraduates at Harvard has a complaint 
> about grade inflation.  More than 48% of all undergraduate
> grades last year were A.  (In 1986, it was only 34% or so.)
> Only 6% or present grades were C or D or F.
> 
> The dean has asked the faculty to discuss it, which is
> as much as she can do.  I don't know: Would the A's 
> emerge as scores on-a-curve, or are the lessons so
> easy that all the answers are right?
[ snip, rest]
Section A  of the  NY Times on Wed., Dec 5, had another 
article (page 14) and a column (page 21).

There were specific comments *contrary*  to some obvious notions
of grade "inflation" as an arbitrary and bad thing:  some were
presented as opinion, and other as apparent fact.  Recent Harvard
students have higher SATs than ever.  Students at a particular level
(of SAT, or otherwise) supposedly are performing better.  
The Dean of Harvard College (a subunit, I think)  says that his
students (in computer science) handle some previously-tough 
problems much more easily.  [ And I wonder, Is that peculiar to cs.]
Someone else was quoted,  that the performance needed for an A 
had not changed.

Amongst the commentary in the column -
Comments on educational research:  Good students (some 
research says) learn more if top grades are kept lower, but 
lower grading can discourage poorer students and increase 
dropout rates.   - Both effects are easy to imagine, somewhere,
sometime, 

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Important notice about Final Exam on Thursday

2001-12-05 Thread Rich Ulrich

I have a question concerning one detail -

On 5 Dec 2001 10:04:38 -0800, [EMAIL PROTECTED] (Dennis Leitner) wrote:
 
> [ ... ]  To use multiple regression
> you need to code the independent variables:  Gender has 2 level; so
> there needs to be 1 coded vector.  SES has 3 levels so you need 2 coded
> vectors.  And then there will be 2 coded vectors for the interaction.  A
> total of 5 vectors.  So far so good.  Most of you did that.  But then to
> test for the Gender Effect you need to get the R-squared (or SS) from a
> FULL MODEL with all 5 vectors and a RESTRICTED MODEL leaving out the
> Gender Variable (so only 4 vectors remain).  And THEN you use THE F
> formula (R-squared full - R-sqaured restricted etc) to calculate the F.
> You cannot read the F from the printout because it is NOT over and
> above.  [ ... ]

I know, I can't be sure, since I don't have any version of the  
Q and A  or printout  in hand.  But I can't imagine what else is 
being discussed

The F-test on a 1-df  hypothesis  is given in ANOVA or Regression
by  the test on the difference in R-squared two models.

And that is EXACTLY the same as the F-test on the beta in the fuller
model; that was (after all) a partial-beta.  

What am I missing?  


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When does correlation imply causation?

2001-12-05 Thread Rich Ulrich

On 5 Dec 2001 08:52:41 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

> correlation NEVER implies causation ...
That is true
 - in the strong sense, and
 - in formal logic, and
 - as a famous quotation among researchers.
(And, reported as wrongly contrasted to 'ANOVA'.)

Or, correlation always implies causation
 - intuitively, or
 - naively, or
 - in the weak sense.

(I admit it -- I see a sexy correlation and 
'implications'  cross  my mind.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What usually should be done with missing values ...

2001-12-05 Thread Rich Ulrich

On Tue, 04 Dec 2001 17:39:53 GMT, Jerry Dallal
<[EMAIL PROTECTED]> wrote:

> Rich Ulrich wrote:
> > 
[ ... ]

> > I don't see much difference.  "Identifying predictors" by regression
> > analyses -- what is that advice supposed to mean?   The criticisms
> > of stepwise selection say that it gives you the wrong variables, not
> > 'merely' (as if that were trivial)  the wrong weights.
> > 
> > Am I missing something?  (I am not totally against stepwise;
> > just, mostly.)

JD >
> There are two issues: determining the right set of variables and
> predicting the response.  Stepwise can be deadly for the former
> without, I believe, being too bad for the latter.  I'm willing to
> recant if someone with authority claims otherwise.

I try to avoid the word "predict"  when I am fitting parameters
to describe a set of data or retrieve a formula - "post-dicting."

That's okay, if that is what you mean.  You are just talking 
about arriving at the minimum squared error in the fit.  Oh,
someone may ask, why not use them all?  In my experience,
I knew that there were supposed to be a certain number of
variables in the equation; so I hoped to recover the actual 
formula, by restricting the variables.

Starting out with pretty good hints as to variables and 
transformations, and some exact, tabled-up, results, 
I once used stepwise selection to recover a formula 
for the wind-chill index.  Another time, I recovered the NFL
formula for rating quarterbacks (and detected, with pretty
high confidence, an error in the raw numbers).

Right ballpark?

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: KS Statistic and adjustments

2001-12-04 Thread Rich Ulrich

On Tue, 4 Dec 2001 15:57:35 + (UTC), "Steve Jones"
<[EMAIL PROTECTED]> wrote:

> Hi everyone,
> I'd like your opinion on the following.
> I am interested in computing the KS Statistic.
> I have a binary choice variable --Buy or Not Buy
> (coded 1 or 0). Using logistic regression I
> have calculated the probability of buying.

Kolmogorov-Smirnov?
You won't do that on a binary variable.

What are you planning to use it for?  
Having a scaling that is rational is usually more
important than having one that is strictly normal;
and where it matters, anyway, is for the tests.

But you should be aware that you do *not*  rely
on the simple, overall ANOVA tests that you generate 
after "weighting" to match a population.  -- If the 
weighting is not so slight as to make no practical
difference, you have to use programs like SUDAAN 
for testing  the contrasts.

> 
> In preparing the data for regression, I did
> a sort of stratified sampling, and generated
> a weight variable (W) that will be applied to
> the sample so that the sample will resemble
> the true population distribution.
> The question: After already included the
> weight in computing the probability of buying,
> is it necessary to use the weight again in
> calculating the KS stat.
> 
> Any arguments for or against?

If you want the KS stat as a TEST about your sample, 
you don't want weighting.  If you want an empirical,
descriptive measure, maybe you would weight -- but
WHO ever used the KS  as a descriptive measure?
I don't remember that.

I think you are over-valuing normality and over-valuing
the  testing of normality.  Also, KS  is not particularly
great for  'pre-testing for normality'.  Assuming that is
your purpose, I think Wilks-Shapiro is a better test.  And 
a box-and-whisker plot shows more about bad outliers
(ones that you need to make decisions about).

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What usually should be done with missing values ...

2001-12-04 Thread Rich Ulrich

On Mon, 03 Dec 2001 20:57:33 GMT, Jerry Dallal
<[EMAIL PROTECTED]> wrote:

> Rich Ulrich wrote:
> > 
> > On 2 Dec 2001 16:49:25 -0800, [EMAIL PROTECTED] (Kevin C. Heslin)
> > wrote:
> > 
> > > Jenny -- here's a way to impute continuous variables using SAS:
> > >
> > > Regression analysis is performed on a continuous variable until significant
> > > predictors of the continuous variable are identified.  The parameter estimates

RU> > 
> >  - "until significant predictors of the continuous variable are
> > identified."
> > 
> > That does limit the exercise to being  strictly exploratory.
> > 
> > You can see my stats-FAQ  (or groups.google search)
> > for criticisms of stepwise selection.

JD> 
> This is a bit different, though.  The criticisms of stepwise
> selection are directed toward the attempt to assess the contribution
> of individual predictors.  The predictions themselves should be

I don't see much difference.  "Identifying predictors" by regression
analyses -- what is that advice supposed to mean?   The criticisms
of stepwise selection say that it gives you the wrong variables, not
'merely' (as if that were trivial)  the wrong weights.  

Am I missing something?  (I am not totally against stepwise; 
just, mostly.)
 

> okay, except in this instance the subset of cases on which the
> predictions are based will vary according to the set of variables
> that is being considered. Hurts my head to think about it, but I
> wouldn't be surprised if someone else said 'not to worry' if
> "missing at random" is appropriate.

People most often ought to drop variables that are highly missing, 
for reasons that are both logical and statistical.  

Sometimes you can salvage a question by recoding, for instance,
< Yes/ No/ Missing>   to  <(definite) Yes/ Other >.  

Also, you can salvage a question by using two variables
for test purposes:  one that dummy-codes Missing as Yes/No;  
plus the original variable, with an arbitrary score for the Missing.
You then consider the two variables together.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What usually should be done with missing values ...

2001-12-03 Thread Rich Ulrich

On 2 Dec 2001 16:49:25 -0800, [EMAIL PROTECTED] (Kevin C. Heslin)
wrote:

> Jenny -- here's a way to impute continuous variables using SAS:
> 
> Regression analysis is performed on a continuous variable until significant
> predictors of the continuous variable are identified.  The parameter estimates

 - "until significant predictors of the continuous variable are
identified."

That does limit the exercise to being  strictly exploratory.

You can see my stats-FAQ  (or groups.google search) 
for criticisms of stepwise selection.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students: A Statistical Perspective

2001-12-02 Thread Rich Ulrich

 - I guess I am commenting on the statistical perspective, 
at least, to start with.

On Fri, 23 Nov 2001 16:22:46 GMT, "L.C." <[EMAIL PROTECTED]>
wrote:

> The question got me thinking about this problem as a
> multiple comparison problem. Exam scores are typically
> sums of problem scores. The problem scores may be
> thought of as random variables. By the central limit theorem,
> the distribution of a large number of test scores should look
> like a Normal distribution, and it typically (though not always)
> does. Hence the well known bell curve. (Assume, for the sake
> of argument that it holds here.)
> 

The tests that I took throughout elementary and high school
(grades 1-12)   were certainly not 'normally distributed'  or 
my scores would have had variation.  Tests or homework: 
everything was truncated at the top.  

Consistency seems to be demanded by the system of 
putting in a 90-point cutoff for A.   If the average were 
75 or 80, a top score of 100  has 1/3 or 1/4  the effect 
of tossing in a bottom score of 0.  (This does seem 
more appropriate as a 'meta-lesson'  at the elementary
level, that it would be at the college level.  That is, I 
can see where it can be more appropriate to judge college 
students by best-performance, or end-of-term performance,
where, for el-hi  students, doing nothing but preparatory 
work, the *consistency*  is part of the lesson.)

I suppose that elementary textbooks *might*  be calibrated
so that hitting 90% of the answers correct is what ought to
justify an A.  - In that case, my teachers *might*  have been
judged by whether they assigned the easier questions or
the tougher ones.  The teachers did use some discretion
in what they assigned.

I do believe that my teachers did not put many really tough
questions on a test or in homework BECAUSE   that would
'bring down the average' -- and they did not have any other
accepted model to use except the 90-point-A model, and
they wanted to give a 'proper'  proportion of A's.   As a 
consequence, I arrived at college without ever learning to 
study.  No course had ever placed any demand on me.

The usual notion of 'grading on the curve'  before college
was that a teacher would give more A's  than the 90-point 
model would justify; so we students tended to want it.  
But I don't remember it happening much at all.

I first matriculated at Rice where the freshman classes were
huge, and were pre-announced as being 'on the curve' --
though the cutoffs were not absolutely fixed in advance.
After making test scores like 35 and 4 (out of 100), I could
*learn*  from the scoring.  I don't think I would have learned
to 'show my work'  -- at least, not so quickly and  dramatically --
if I had not had that 4 on a physics test.

With the curve, and low, low averages, you do notice 
that a single *good*  performance can outweigh several
poor ones.  So that is good.

Now, in the NY Times, just a week or two ago.  The
dean of undergraduates at Harvard has a complaint 
about grade inflation.  More than 48% of all undergraduate
grades last year were A.  (In 1986, it was only 34% or so.)
Only 6% or present grades were C or D or F.

The dean has asked the faculty to discuss it, which is
as much as she can do.  I don't know: Would the A's 
emerge as scores on-a-curve, or are the lessons so
easy that all the answers are right?

What is the purpose of an A, anyway?  There are
relatively few courses where the grade is matched against
an outside criterion.  When I think about it, I don't think
the existence of the outside criterion is apt to matter.
Law school -- state exams are a breeze in a few states, 
or they can be almost impossible.  Different schools 
in the state will have different standards, I am sure.

Medical boards, on the other hand, are national, and 
I think they are usually passed by decent students.
But I don't think of the MDs I have known as being 
obsessed about the college boards that they, personally,
passed some years ago.

I think the professors in either sort of school have standards
from elsewhere for the grades they assign.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Interpreting p-value = .99

2001-12-02 Thread Rich Ulrich

On Sat, 1 Dec 2001 08:20:45 -0500, [EMAIL PROTECTED] (Stan Brown)
wrote:

> [cc'd to previous poster]
> 
> Rich Ulrich <[EMAIL PROTECTED]> wrote in sci.stat.edu:
> >I think I could not blame students for floundering about on this one.
> >
> >On Thu, 29 Nov 2001 14:39:35 -0500, [EMAIL PROTECTED] (Stan Brown)
> >wrote:
> >> "The manufacturer of a patent medicine claims that it is 90% 
> >> effective(*) in relieving an allergy for a period of 8 hours. In a 
> >> sample of 200 people who had the allergy, the medicine provided 
> >> relief for 170 people. Determine whether the manufacturer's claim 
> >> was legitimate, to the 0.01 significance level."
> 
> >I have never asked that as a question in statistics, and 
> >it does not have an automatic, idiomatic translation to what I ask.
> 
> How would you have phrased the question, then? Though I took this 
> one from a book, I'm always looking to improve the phrasing of 
> questions I set in quizzes and exams.
 [ snip, rest]

"In a LATER sample of 200 ... relief for ONLY  170 people."

The Query you give after that should not pretend to be ultimate.
Are you willing to ask the students to contemplate that
the new experiment could differ drastically from the original 
sample and its conditions?
 
"Is this result consistent with the manufacturer's claim?"
 - you might notice that this sounds  'weasel-worded.'

Well, extremely-weasel-worded  *ought*  to be suiting,
for  *proper*  statistical claims from non-randomized trials.  
For the example: I would expect 15%  of a grab-sample
being treated for 'allergy'  would actually have flu or a cold.
Maybe the actual experiment was more sophisticated?

"What do you say about this result? (include a statistical
test using a nominal alpha=.01)."

Also,  "Why do I include the word 'nominal'  here?"  
Ans:  It means 'tabled value'  and it helps to emphasize
that it is hard to frame a non-random trial as a test;  the
problem is not presented with any such framing.

Hope this seems reasonable.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: (none)

2001-11-30 Thread Rich Ulrich

On 30 Nov 2001 07:34:14 -0800, [EMAIL PROTECTED] (Carl Huberty)
wrote:

  "Need help.  Would someone please give me a reference that 
discusses a chance value of a canonical correlation coefficient 
(or the square thereof).  That is my question is:  What is the
expected value of R when rho is zero?"

Have you tried the 2nd Edition of Cohen's book (1989 or 90)
on power analysis?  (The 1st and Revised editions
don't have the final chapter or two.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Interpreting p-value = .99

2001-11-30 Thread Rich Ulrich

I think I could not blame students for floundering about on this one.

On Thu, 29 Nov 2001 14:39:35 -0500, [EMAIL PROTECTED] (Stan Brown)
wrote:

> On a quiz, I set the following problem to my statistics class:
> 
> "The manufacturer of a patent medicine claims that it is 90% 
> effective(*) in relieving an allergy for a period of 8 hours. In a 
> sample of 200 people who had the allergy, the medicine provided 
> relief for 170 people. Determine whether the manufacturer's claim 
> was legitimate, to the 0.01 significance level."
> 
> (The problem was adapted from Spiegel and Stevens, /Schaum's
> Outline: Statistics/, problem 10.6.)
[ snip, rest ]

"Determine whether the manufacturer's claim was legitimate,
to the 0.01 significance level."  

I have never asked that as a question in statistics, and 
it does not have an automatic, idiomatic translation to what I ask.


I can expect that it means, "Use a 1% test."  But, for what?

After I notice that the outcome was poorer than the claim, 
then I wonder if the test is, "Are these data consistent with
the Claim? or do they tend to disprove it?"  That seems 
some distance from the tougher, philosophical question of 
whether, at the time it was made, the claim was legitimate.

That claim could NEVER, legitimately, have been *based*  
on these data.   That is an idea that tries to intrude itself,
to me, and makes it difficult to address the intended question.


 - By the way, it also bothers me that "90% effective"  is 
apparently translated as "effective for 90% of the people."
I wondered if the asterisk was supposed to represent "[sic]".

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: N.Y. Times: Statistics, a Tool for Life, Is Getting Short Shrift

2001-11-30 Thread Rich Ulrich

On Fri, 30 Nov 2001 14:38:33 GMT, mackeral@remove~this~first~yahoo.com
(J. Williams) wrote:

> On 29 Nov 2001 07:03:13 -0800, [EMAIL PROTECTED] (Robert J.
> MacG. Dawson) wrote:
> 
> 
> >There is probably a reverse trend in the extreme tail; people probably
> >overestimate the probability of getting (say) red fifty times in a row
> >at Roulette simply because we don't have a good feel for really large
> >and small numbers. 
> 
> I think you are right in that assumption.  When I taught probability,
> I found students had difficulty sensing  numerical enormity or its
> opposite in scientific notation or lots of zeros.  Dealing with 16
> zeros to the right of the decimal, for example,  becomes a complete
> abstraction.  

 - whereas, by contrast, we scientists can right it out with
"scientific notation"  with its powers of ten, and have something 
concrete, not abstract, because it is additive in the exponents....
or am I just making another complete abstraction of it?

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: t vs. z - recapitulation

2001-11-29 Thread Rich Ulrich

 - I am just taking a couple of questions in this note -

On Thu, 29 Nov 2001 13:16:24 +0100, "Gaj Vidmar"
<[EMAIL PROTECTED]> wrote:
[ ... ]

I saw some decent comments about the table;  the table
was not very useful.

z is used with large N  as 'sufficiently good' approximation for  t. 

z  is used when "variance is known" -- which is, in particular, 
when statistics are computed that were based on dichotomies
or on ranks (if there were no ties).  That's basically when 
variances are known.  With big samples and ties, you are 
probably better off doing your rank-transformation, then using
the F-test or t-test.

Depending on the outliers, ANOVA (t-tests) might be useless,
regardless of how big the sample *might* get -- that happens
more often than careless analysts expect, when they don't 
watch for outliers.  -- If  you can't opt for a parametric transform,
you might need to test after rescoring as 'ranks'  or into categories
(two or more).


> 
> Note 2: t-test is very robust (BTW, is Boneau, 1960, Psychological Bulletin
 - not in *my* opinion (see below) -

> vol. 57, referenced and summarised in Quinn and McNemar, Psychological
> Statistics, 4th ed. 1969, with the nice introduction "Boneau, with the
> indispesable help of an electronic computer, ...", still an adequate
> reference?), whereby:
> - skewness, even extreme, is not a big problem
> - two-tailed testing increases robusteness

 - I was annoyed when I learned that those old-line authors 
would decide that a test was 'robust with two-tails'  when it 
rejected 9%  in one tail in 1% in the other.  It felt somewhat 
like having been lied-to.  I still disagree with that opinion.

Fortunately, the *problem*  of really-bad-p-values (in both
directions)  does not exist for equal N.  Unfortunately, even
for equal N, there *can*  be a great loss in statistical power.
So, you should be unhappy to see great skewness.

But for either equal or unequal N, I am much happier if I can
trust that the variable has been transformed to its proper
metric; and if that metric does not have skewness, or 
heterogeneity of variance.

If a variable *needs*  to be transformed, please transform it.
(But the 'need' issue is worth its own discussion.)

> - unequal variances are a serious problem with unequal N's with larger
> variance of smaller sample

Oh, a problem with larger variance in EITHER sample.
A different problem, one way or the other.

Outliers cause loss of power for ANOVA, just as much as 
outliers screw up a mean -- If you see outliers, ask, Are you 
sure ANOVA is the right tool?  

> 
> Now, what to do if t is inadequate? - This is a whole complex issue in
> itself, so just a few thoughts:
> - in case of extreme skewness, Mann-Whitney is not a good alternative
> (assumes symmetric distrib.), right?
 [ ... ]

It assumes *same*  distributions in two samples, not necessarily
symmetric.  What is your hypothesis going to be?  What can you
fairly conclude, if one sample occupies both ends of the distribution?

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Normal distribution

2001-11-29 Thread Rich Ulrich

On Thu, 29 Nov 2001 14:37:14 -0400, Gus Gassmann
<[EMAIL PROTECTED]> wrote:

> Rich Ulrich wrote:
> 
> > On Thu, 29 Nov 2001 15:48:48 +0300, Ludovic Duponchel
> > <[EMAIL PROTECTED]> wrote:
> >
> > > If x values have a normal distribution, is there a normal distribution
> > > for x^2 ?
> >
> > If z is standard normal [ that is, mean 0, variance 1.0 ]
> > then z^2  is chi squared with 1 degree of freedom.
> >
> > And the sum of S   independent  z  variates
> > is chi squared with S degrees of freedom.
> 
> Hold it! The sum of S independent z variates is normal.
> The sum of the _squares_ of S independent z variates is
> chi squared with S degrees of freedom.
> 
> (But I am sure you knew that.)

oops- make that < z^2 >  for < z >  of course -


-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Normal distribution

2001-11-29 Thread Rich Ulrich

On Thu, 29 Nov 2001 15:48:48 +0300, Ludovic Duponchel
<[EMAIL PROTECTED]> wrote:

> If x values have a normal distribution, is there a normal distribution
> for x^2 ?

If z is standard normal [ that is, mean 0, variance 1.0 ]
then z^2  is chi squared with 1 degree of freedom.

And the sum of S   independent  z  variates
is chi squared with S degrees of freedom.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: biostatistics careers

2001-11-26 Thread Rich Ulrich

On 19 Nov 2001 20:56:58 GMT, [EMAIL PROTECTED] (A.J. Rossini)
wrote:

> >>>>> "RS" == Richard Seymann <[EMAIL PROTECTED]> writes:
> 
> RS> And if I may muddy the waters even more, what is the
> RS> difference between biostatistics and biometry?  Dick
> 
> Depends on which definition of "biometry" you are using.  One
> definition used to be an older name for what is now biostatistics (and
> what might again become biometry, according to how some in the field
> want to rename it again to make it more relevant to measurement and
> design).  That's a nice murky response, which shows how our language
> is ill-defined.
> 
> There are other definitions, which have little to do with statistics. 

I was a little startled by my google search.  "Definition of
biometry"  gets the Cornell Department of Biometrics, and a 
couple of other things, among only 6 hits.

"Definition of biometrics" returns dozens of references to 
security systems, fingerprints and retinal scans, etc.  That
is a definition that seems to be winning out, owing to its 
new, computerized, commercial potential.

There were no hits that would make me think of the
journal named Biometrika, which I remember as being 
highly mathematical.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: best inference

2001-11-21 Thread Rich Ulrich

On Wed, 21 Nov 2001 10:24:54 -0600, Bill Jefferys
<[EMAIL PROTECTED]> wrote:

> In article <[EMAIL PROTECTED]>, 
> [EMAIL PROTECTED] (Dennis Roberts) wrote:
> 
> #on this near holiday ... at least in the usa ... i wonder if you might 
> #consider for a moment:
> #
> #what is the SINGLE most valuable concept/procedure/skill (just one!) ... 
> #that you would think is most important  when it comes to passing along to 
> #students studying "inferential statistics"
> #
> #what i am mainly looking for would be answers like:
> #
> #the notion of 
> #
> #being able to do __

> 
> I'd say "the notion of inverse probability", but of course that's 
> because of where I am coming from :-)
> 

I'd say, being able to re-frame the researcher's 
fuzzy curiosity as a "testable hypothesis"  -- I think
*that*  is where you get the probability that is to be inverted.
Good hypothesizing would not be so valuable except 
that it is  rare.  And there are so many people who are 
close to it, and need it.  So it ought to be valuable, as
a commodity in demand.

As a commodity, at least, "the notion of inverse 
probability"  is less valuable because it is out of reach.  
It is of concern, I think, to the people who have
finally achieved a p-value, and wonder what to do next.

(Actually, I don't have a quick opinion, but I thought
Promoting Tests was a good way to pull Dennis's leg.)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When Can We Really Use CLT & Student t

2001-11-21 Thread Rich Ulrich

On 21 Nov 2001 10:18:01 -0800, [EMAIL PROTECTED] (Ronny
Richardson) wrote:

> As I understand it, the Central Limit Theorem (CLT) guarantees that the
> distribution of sample means is normally distributed regardless of the
> distribution of the underlying data as long as the sample size is large
> enough and the population standard deviation is known.
> 
> It seems to me that most statistics books I see over optimistically invoke
> the CLT not when n is over 30 and the population standard deviation is
> known but anytime n is over 30. This seems inappropriate to me or am I
> overlooking something?
[ snip, rest ]

It seems to me that you have doubts which *might* be justifiable.

Do you have a professor who is prone to glib generalizations?
Do you have a lousy text?

I do wonder if your textbooks actually say what you accuse them of, 
or if you are guilty of hasty overgeneralization.  I have scanned 
textbooks in search of  errors like those, but I hardly ever find any.
Gross mis-statements tend to be in  "handbooks"  and in 
(unfortunate) interpretative  articles by non-statisticians.

(Can you cite "chapter and verse"?)

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: biostatistics careers

2001-11-19 Thread Rich Ulrich

On Mon, 19 Nov 2001 08:07:33 -0500, Bruce Weaver
<[EMAIL PROTECTED]> wrote:

> 
> On Sun, 18 Nov 2001, Stan Brown wrote:
> 
> > What _is_ "biostatistics", anyway? A student asked me, and I
> > realized I have only a vague idea.
> 
> 
> There was a thread on "biostatistics versus statistics" a couple years
> ago, I think, but I was unable to find it at google groups.  Maybe someone
> out there saved some of it.

groups.google.com  on < biostatistics statistics > 
 - I found a couple of notes, within the top 100.  
There were several comments on 23 Feb 2000  
with the subject line, "re:biostatistics".  They mention
that medical background is important.  And vocabulary.

Also, as I vaguely remembered, I personally had answered a 
similar question, on 18 Feb 1998:

=== from my 1998 comment:
 - There are a couple of dozen or so U.S. universities that include
a "Graduate School of Public Health."  Here at the University of
Pittsburgh, it is the GSPH  that awards a degree in biostatistics.

The course work for the degree does include courses which would not
be required for "statistics" as I imagine it -  epidemiology (chronic
vs  acute diseases), vital statistics, health services administration.
[ ... ]
=== end of 1998 citation

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: [HELP NEEDED] What is the best technique to analyze the following experiment?

2001-11-18 Thread Rich Ulrich

On 16 Nov 2001 09:34:52 -0800, [EMAIL PROTECTED] (S.) wrote:

> I performed the following experiment:
> 
> Each user (U) used several interfaces (I). Both U and I are to be
> treated as random factors. For each U and I combination, time (T),
> errors (E) and satisfaction (S) were measured. The data looks
> something like:
> 
> U I T E S
> ---   ---   ---   ---   ---
> U1I1100   1090
> U1I2200   2080
> U1I3300   3070
> U1I4400   4060
> U2I1102   1191
> U2I2198   1881
> U2I5500   5050
> U2I6600   6040
> .
> .
> .
> etc.
> 
> Please note that NOT all the users used all the interfaces.
> 
> The question is: I wish to find the correlations between T, E and S
> (viz., nullify the effects of U and I). What is the best statistical
> method of doing this? I think something along the lines of Anova or

For the little bit of data shown, the variable *I*  has a huge effect,
with R-squared of maybe 0.99 with each of the three variables, T, E
and S, and that happens while I call it a continuous variable.  
So it would be just as important, with a waste of degrees of freedom,
if it is used as categories; the table shows it coded as categories,
I-1 to I-6.  

High R-squared puts you into the situation where subtle choices
of model can make a difference.  Is it appropriate to remove the
effect of *I*  by subtraction, or by division? - by category, or by 
treating it as continuous?

> Variance Components should do that trick... I have SPSS, so any advice
> on how to interpret the output will be most appreciated (please bear
> in mind that I do not have a degree in statistics).

If it is strictly correlation that you want, you can ask for the 
intercorrelations, while partial ling out the U and I variables.
If *I*  and U  are to be partialled-out as categories, you can create
a set of dummy variables, and partial-out those.  

The result that you get will  *not*  be robust against scaling
variations (linear versus multiplicative, for instance).  That is
a consequence of the high R-squared and the range of numbers
that you have.  I suspect that the observed R-squared values
might vary in a major way if you just change the raw data by a few
points, too -- Note that prediction with an R-squared of 0.99  
has *twice*  the error of an R-squared of 0.995, and so on; 
that is approximately the same as the difference between 0.1 
and 0.2, in certain, practical consequences.

If it will please you to reduce the eventual intercorrelations to
zero, a proper strategy *might*  be to try alternative models to 
see if you can produce that result.

Of course, in practice, it should be a great deal of help
to know what the variables actually, are, and how they
are scored, etc., to know what transformations are logical
and appropriate.  I suspect that data, as stated, leave out
some conventional standardization, and so the observed
correlations are mainly artifacts.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Usability of skills and knowledge

2001-11-16 Thread Rich Ulrich

On 16 Nov 2001 09:16:08 -0800, [EMAIL PROTECTED] (Rolf Dalin) wrote:

> In a discussion about the desired direction of development of intro 
> level statistics courses, a group where I am a member came to a 
> preliminary agreement that It is important to develop applied 
> statistics. I started to think about that concept in terms of main goals 
> of a course. The two main goals I suggest are 
> 
> 1. abilities to use statistics in scientific work
> 2. ability to study statistics further
> 
> So nr. 1 is the goal that concerns the "applied" part. So I went one 
> step further to try to express aspects of this, and I call it the 
> USABILITY aspects of the introductory course. I know that this word 

Relating to the Subject line, usability of skills:

 - an important job-skill:  to not-screw-things up while getting
something done.  
 - important job-skills:  to do  right.
 - important skill:  self-assessment, to know if the task *can* 
be done, or *has*  been done pretty-much correctly and reliably.

When it comes to having "basic skills in statistics,"  knowing 
the basic vocabulary is vital.  

However, the idea keeps coming back to me that 
people misuse the basic statistics because they don't 
understand the basics of logical and scientific inference.
So, they need to re-study plane geometry.  And read up
on philosophy of science, and experimental design.
Or join a debating society?

> is used in the context of evaluating web applications. Now to my 
> questions: Can anybody direct me to an article or book that 
> discusses usability oin the context of knowledge? Is there a better 
> direction to follow when discussing the usefulness/applicability of 
> statistics skills and knowledge?

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Fw: We need your help!!!

2001-11-13 Thread Rich Ulrich

On 12 Nov 2001 11:41:45 -0800, [EMAIL PROTECTED] (Carl Huberty)
wrote:

> It would be greatly appreciated if I could get references for the six topics
> mentioned in the message below.  I assume that Conover (1999) discusses the
> first topic.  But beyond that I am at a loss.  Thanks in advance.
> 
> Carl Huberty
> - Original Message -
> From: <[EMAIL PROTECTED]>
> To: <[EMAIL PROTECTED]>
> Cc: <[EMAIL PROTECTED]>
> Sent: Sunday, November 11, 2001 9:10 AM
> Subject: We need your help!!!
> 
> 
> > Dear Sincere Prof. Huberty
> > I hope that you are spending a happy time .
> > I have a sincere friend from Egypt. He is going to prepare for his
> > proposal and he is going to make his defense in the next few days. He
> > asked our help to provide him with some information about:
> >
> > Equations used to make Non-parametric factorial
> > ANOVA :-
> > Bradley's collapsed and reduce test.
> > Harwell-Serlin's L  test.
> > Blair-Sawilowsky's adjusted rank transform test.
> > Puri-Sen test.
> > Fawcett and Salter's aligned rank test.
> >

A few minutes of searching with google shows that these names
and tests are not widely known.   Several are mentioned
in the same articles.  It looks like someone needs the 
Proceedings of the 
Joint Statistical Meetings in Toronto, August 15-18, 1994.


 Harwell-Serlin's L  test.
http://seamonkey.ed.asu.edu/~behrens/edstat/newsletter_95.html

 Blair-Sawilowsky's adjusted rank transform test.   Also
Harwell-Serlin:
http://lib.stat.cmu.edu/joint94/section-a.html
http://lib.stat.cmu.edu/joint94/Abstracts/0486

 Puri-Sen test.  
http://www.iuinfo.indiana.edu/HomePages/021999/text/quantifying.htm

 Fawcett and Salter's aligned rank test.
see 0486, above.

Several tests are mentioned in a poster session with abstracts at
http://www.stat.ufl.edu/symposium/1999/nonpar/poster-abstracts.html

"The Best Test for Interaction in Factorial ANOVA and ANCOVA 
Todd C. Headrick and Shlomo S. Sawilowsky 
Wayne State University"

I wonder...  Isn't an education, these days, supposed
to develop the ability to go online and search for topics like these?

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Standard Deviation!

2001-11-12 Thread Rich Ulrich

On Sun, 11 Nov 2001 01:30:27 +1100, "David Muir"
<[EMAIL PROTECTED]> wrote:

> Presently the Gaming Industry of Australia is attempting to define various
> new 'definitions of Standard Deviation'...in a concept to define infield
> metrics for the analysis of machines in terms which imply whether a machine

Google - does not find "infield metrics" or ""infield observable"
(below)  and I find the terms baffling.  That is part of what makes
the question baffling to me.  

The "standard deviation" is also a poser.  I think someone will
have to cite the proposed law.  

I do see where regulated gambling and "probability"  raise some
problems.  If slot machines are supposed to return a certain 
percentage by law, is there a control on each machine to assure
that each machine, separately performs within certain limits?

Does this mean that the "winning"  has to be monitored, so that
it can periodically be adjusted?  - that would make the old-style
sucker potentially correct, the one who insisted on keeping  with
their  machine because it was 'coming due.'  

But if the odds are more realistic and independent, then some 
machine, out of a million sequences for all machines, will 
have a long, losing sequence that is a-million-to-one.


> is being operated with respect to its defined percentage or in fact outside
> its defined region - i.e.. Illegally manipulated.
> 
> My understanding of the Standard Deviation metric does not fit the mixed
> (confused) proposals of the industry. Therefore I ask if a suitable
> mathematician may be available to examine the problem.
> 
> The gaming industry seems to want the metric used in terms of a periodic
> infield observable, my feeling is that it is inappropriate and another
> method must be provided if possible. The first part is to confer the
> inappropriate aspect of STDDEV to periodic observations. After which
> appropriate methods are required!
> 
> Here is the problem:
> 
>  Imagine a game with two components, the base game and a feature. During the
> base game prizes from say 20 units to 5000 are awarded (the standard
> deviation being well defined). During  the base games at
> statistically 1 in say 120 occurrence, a feature game occurs which uses the
> same base game prizes though all are multiplied by 3 i.e. 60 to 15000 unit
> prizes.
> 
>  What method should be used to define the standard deviation?
> 
> For anyone able to provide a provable solution a monetary prize of at least
> $250US is available.

   3 times the payoff == 9 times the sum of squares.
So if you know what the "standard deviation"  is for 119
out of 120 occurrences, and you call that *variance*  1.0;  
and you replace variance #120 with 9.0;  then you can 
see how the standard deviation is increased.  But this 
is a silly problem to me because I don't make any sense
out of 'standard deviation.'

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Z Scores and stuff

2001-11-09 Thread Rich Ulrich

On Thu, 8 Nov 2001 18:31:57 +, Mark T <[EMAIL PROTECTED]>
wrote:

> Hi,
> 
> What are the formulae for calculating the mean to z, larger proportion and smaller 
>proportion of a z-score (standardised score) on a standard normal distribution? I 
>know about tables listing them all, but I want to know how to work it out for myself 
>:o)

Do you want the calculus, or just a numerical approximation?

For starters, in my stats-FAQ, see

http://www.pitt.edu/~wpilib/statfaq/gaussfaq.html

 - contents contributed by John D'Errico.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Testing for joint probability between 2 variables

2001-11-01 Thread Rich Ulrich

On Tue, 30 Oct 2001 21:10:02 -, "Chia C Chong"
<[EMAIL PROTECTED]> wrote:

[ ... ]
> 
> The observations were numbers. To be specified, the 2 variables are DELAY
> and ANGLE. So, basically I am looking into some raw measurement data
> captured in the real environment and after post-proceesing these data, I
> will have information in these two domains.
> 
> I do not know whether there are linearly correlated or sth else but, by
> physical mechanisms, there should be some kind of correlation between them.
> They are observed over the TIME domain.

I don't think it has been answered yet, whether they are
correlated because they are autocorrelated in a trivial way.
What does it mean here -- or does it happen to signify 
nothing -- that observation is  "over the TIME domain".  

That is, you have a real problem yet to be faced, if these are 
measured as  "cumulative delay"  and "cumulative angle".

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: ranging opines about the range

2001-10-31 Thread Rich Ulrich

[ I have rearranged Zar's note.]   After this one, 

> >>> Harold W Kerster <[EMAIL PROTECTED]> 10/29/01 04:31PM >>>
>   If you define the range as max - min, you get zero, not one.  What 
> definition are you using.

On 29 Oct 2001 16:11:15 -0800, [EMAIL PROTECTED] (Jerrold Zar)
wrote:

> I was referring to the definition that others on the list had proposed: 
> max - min +1.  It is NOT a definition with which I agree. 

If  {max - min + K}  had been a formal proposal, it should have 
stated that K  will be  "1"  when the numbers are reported to
the nearest integer, but -- generally -- K should sensibly
reflect the precision of  measurement-and-reporting.

I don't know who needs it in the real world most of the time,  
but using K  gives a better -- usually safer -- estimate when  
you are using the range to estimate the standard deviation.

But.
Whenever {max-min}   is small enough that K  is a sizable
correction, someone needs to speak carefully about the 'range.'
To put it another way:  If all the cases all are observed at
the same value, and it matters, then the audience really 
does deserve to hear the details. 

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Graphics & CORRESPONDENCE ANALYSIS

2001-10-24 Thread Rich Ulrich

On Mon, 22 Oct 2001 17:38:04 +0200, christian <[EMAIL PROTECTED]>
wrote:

> Hello,
> 
> ...have anybody experience or a good idea
> how i can display the result from the correspondence analysis
> in a better way like spss (i.e. excel ), because the visual performance 
> in my humble opinion is not the best !?

If you are looking for a model to program after --
I liked the display of the old BMDP  procedure, CA, as shown
on page 660 (and 664)  of the 1990 manual (Vol 2).

Here is one entry to a US web-site that has a lot on CA.
http://www.okstate.edu/artsci/botany/ordinate/glossary.htm

It has been my impression (from google) that CA is more popular 
in European journals than in the US, so there might be better
sites out there in a language I don't read.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Standardized Confidence Intervals

2001-10-21 Thread Rich Ulrich

On 15 Oct 2001 07:44:33 -0700, [EMAIL PROTECTED] (Warren) wrote:

> Dear group,
> It seems to me that the one issue here is that when we
> measure something, then that measure should have some
> meaning that is relevant to the study hypotheses.  
> And that meaning should be interpretable so that the width 
> of the CI does have meaning...why would you want to estimate 
> the mean if it is "meaningless"?

This reminds me that data analysts sometimes can help 
to make outcomes more transparent.  "Scaled scores"  
of Likert-like items are intelligible when presented as Item-averages,
instead of being Scale-totals for varying numbers of items.

You can describe the relevant anchors for 2.5  versus 3.0,
for group contrasts, change scores, or contrasts between
different scales -- and not get into confusion of how many 
items were added for each total.

I do think that standardized outcomes are usually appropriate
to communicate the size of the outcome effect -- 
even when the measure is fairly well known.

This discussion leads me to conclude that you absolutely
need to describe your sample *more*  thoroughly  
when you are stuck with using standardized measures 
to describe all your outcomes.  How variable is this group?  
How extreme is this group and sample (if it is a clinical trial)? 
A small N is  especially problematic, since you do want 
to show how narrowly (or not) it was selected.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Transformation function for proportions

2001-10-19 Thread Rich Ulrich

On Fri, 19 Oct 2001 11:58:18 +1000, "Glen Barnett"
<[EMAIL PROTECTED]> wrote:

> 
> Rich Strauss <[EMAIL PROTECTED]> wrote in message
> news:[EMAIL PROTECTED]...
> > However, the arcsin transformation is for proportions (with fixed
> 
> It's also designed for stabilising variance rather than specifically inducing
> symmetry.
>  Does it actually produce symmetry as well?
> 
> > denominator), not for ratios (with variable denominator).  The "proportion
> > of sentences in a number of texts that belong to a certain category" sounds
> > like a problem in ratios, since the total number of sentences undoubtedly
> > vary among texts.  Log transformations work well because they linearize
> > such ratios.
> 
> Additionally for small proportions logs are close to logits, so logs are
> sometimes helpful even if the data really are proportions. Logs also go 
> some way to reducing the skewness and stabilising the variance, though 
> they don't stabilise it as well as the arcsin square root that's specifically 
> designed for it.

The transformation is okay but not great for proportions 
less than (say) 5%.   Jez Hill followed up on a reference that
gave him this summer, and posted further detail --
====== from June 27, 2001, Jez Hill.
Subject: Re: [Q ] transforming binomial proportions 
Newsgroups: sci.stat.math

Rich Ulrich wrote in article
[EMAIL PROTECTED]:

> The fixed variance was the main appeal of the approximation,
> "arcsin(sqrt(p))". [snip]
> "A more accurate transformation for small n has been tabulated by
> Mosteller and Youtz."[ Biometrika 48 (1961):433.]

Thanks very much for that - it looks pretty good to me at n=500,
6<=np<=494

FYI: Following up on your reference, Mosteller and Youtz give the
following formula from Freeman and Tukey [Ann. Math. Statist.
21(1950):607].

arcsin(sqrt( np/(n+1) ))/2 + arcsin(sqrt( (np+1)/(n+1) ))/2

which gives asymptotic variance 821/(n+0.5) "for a substantial range
of p if  n is not too small". I find that the improvement is quite
significant, to the point where I would be quite happy to use it even
for np=1, 2 or 3 at  n=500, minor glitches in that region
notwithstanding.
=== end of post
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Transformation function for proportions

2001-10-18 Thread Rich Ulrich

On Wed, 17 Oct 2001 15:50:35 +0200, Tobias Richter
<[EMAIL PROTECTED]> wrote:

> 
> 
> We have collected variables that represent proportions (i. e., the
> proportion of sentences in a number of texts that belong to a certain
> category). The distributions of these variables are highly skewed (the
> proportions  for most of the texts are zero or rather low). So my

Low proportions, and a lot at  zero? 
 
There is no way you can transform to "symmetry"  when 
there is a clump at one end and a long tail at the other.

First thought:  the dichotomy of None/Some  sometimes
contains most of the information that is useful.  Dummy Var1.

Related thought:  "none"  is sometimes a separate dimension
from what is implicitly measured by the continuous values above zero.
If that dimension does seem useful:  Dummy Var2.


> question is: Is there a function that transforms the proportions into
> symmetrically distributed variables? And is there a reliable statistics
> text that discusses such transformations?

"Symmetry"  might happen, and it is good to have for the sake
of testing.  However, describing a scientific  model  with 
meaningful parameters  is a better starting point, and you 
can devise tests from there.   

I mean: it is useful if you have a "Poisson model with a 
Poisson parameter", say, at the stage of setting up a model.  
You might want to take the square root before you do testing, 
and you know that is appropriate for the Poisson;  but the raw 
Poisson parameter is a number that is ordinarily additive.

I have not seen many texts that tackle transformations in
the abstract.  Finney's classic text on bioassay has a few pages.
Or, I think, Mosteller and Tukey, "Data Analysis and Regression."

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Help for DL students in doing assignments

2001-10-15 Thread Rich Ulrich

 - Mathematical logic is a lot tougher when your
translation fails to properly distinguish "any" and
"every"  and "some."


On 15 Oct 2001 07:18:43 -0700, [EMAIL PROTECTED] (Dr. Fairman) wrote:

[ ... ]

> 4.If you agree with item #3 (if not - please argue -  why), it means that
> you are also agree with the statement:
> "every even is (in particular) sum of any two primes".
> That's what you needed me to prove.

Needed:
"Every even number can be written as the sum of two primes" or
For [each] even number  S, there always exists two prime numbers
K and L  such that S= K+L.

Stated by Fairman, 
as read by me (a native speaker of English),
"If K and L are *any*  two primes,  and S is any even number, 
it is true that S = K+L ."  

That is true for arithmetic-modulo-2.  Otherwise, not.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Are parametric assumptions importat ?

2001-10-14 Thread Rich Ulrich

On 12 Oct 2001 11:14:54 -0700, [EMAIL PROTECTED] (Lise DeShea) wrote:

> Re robustness of the between-subjects ANOVA, I obtained permission from Dr. 
> Rand Wilcox to copy three pages from his book, "New Statistical Procedures 
> for the Social Sciences," and place them on a webpage for my students.  He 
> cites research showing that with four groups of 50 observations each and 
> population standard deviations of 4, 1, 1, and 1, the empirical Type I 
> error rate was .088, which is beyond Bradley's liberal limits on sampling 
> variability [.025 to .075].  You can read this excerpt at 

Well, I suggest that a variance difference of 16  to 1 practically
washes out the usual interest in the means.  Isn't that beyond
the pale of the usual illustrations of what is robust?  I may remember
wrong, but it seems to me that Tukey used Monte Carlo with
"10% contamination"  of a sample, where the contaminant
had excessive variances:  10-fold for the variances?

What can you say about an example like that?

Will a Box-Cox transformation equalize the variances? (no?)
Is there a huge outlier or two?  If not -- if the scores are well
scattered -- all the extreme scores in *both*  directions will be 
in that one group.  And a "mean difference"  will implicitly
be determined by the scaling:  That is, if you spread out the 
low scores (say), then the group with big variance will have the
lower mean.


> www.uky.edu/~ldesh2/stats.htm -- look for the link to "Handout on ANOVA, 
> Sept. 19-20, 2001."  Error rates are much worse when sample sizes are 
> unequal and the smaller groups are paired with the larger sigma -- up to an 
> empirical alpha of .309 when six groups, ranging in size from 6 to 25, have 
> sigmas of 4, 1, 1, 1, 1, 1.
> 
> The independent-samples t-test has an inoculation against unequal variances 
> -- make sure you have equal n's of at least 15 per group, and it doesn't 
> matter much what your variances are (Ramsey, 1980, I think).  But the ANOVA 
> doesn't have an inoculation.
> 
> I tell my students that the ANOVA is not robust to violation of the equal 
> variances assumption, but that it's a stupid statistic anyway.  All it can 
> say is either, "These means are equal," or "There's a difference somewhere 
> among these means, but I can't tell you where it is."  I tell them to move 
> along to a good MCP and don't worry about the ANOVA.  Most MCP's don't 
> require a significant F anyway.  And if you have unequal n's, use 
> Games-Howell's MCP to find where the differences are.

Some of us don't like MCPs.  We think that the overall test is
not (or at least, not always) a bad way to start, if a person *really*
can't be more particular about what they want to test.

And if you have unequal Ns, you are stuck with one approximation 
or another, which has to be ugly when the Ns are too unequal; or
else you are stuck with inconsistent statements, where the smaller
difference in means is 'significant' but the larger one is not.
(I am unfamiliar with Games-Howell's MCP.)

Just another opinion.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



  1   2   3   4   5   >