Re: effect size/significance

2001-09-14 Thread Thom Baguley

Mike Granaas wrote:
 I think that we might agree:  I would say that studies need a clear a
 priori rational (theoretical or empirical) prior to being conducted.  It
 is only in that context that effect sizes can become meaningful.  If a

Even then standardized effect sizes may not be very helpful. We need
to know much more information about the effect, the sensitivity of
our measurements and so on.

Thom


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-14 Thread Rolf Dalin

I remember I read somewhere about different effect size measures 
and now I found the spot: A book by Michael Oakes, U. of Sussex, 
Statistical Inference 1990. The measures were (xbar-ybar)/s, 
Proportion misclassified, r squared (biserial corr) and w squared 
(which I think means the same as Rsq adj in ordinary linear 
regression).

I would rather talk about these things as measures of different 
aspects of a relationship between two variables. (A quantitative and a 
qualitative with two categories in Oakes' example.):

Statistical effect, explanatory power and strength of relationship. (... 
even if one could be derived from another ...)

Other aspects which could be added as pieces of information would 
be p-value of test of no relationship, real world effects, causal 
mechanisms, consistency, responsiveness ... (these last from a 
Mosteller and Tukey reference).

If we teach this, I think it would be more obvious that one single 
printout doesnt tell the whole story. And I think it would be a good 
thing to acheeve. Anyway I would be happy to read comments about 
aspects of relationships, since I have only just started to think about 
it in this way.

/Rolf 


 Mike Granaas wrote:
  I think that we might agree:  I would say that studies need a clear a
  priori rational (theoretical or empirical) prior to being conducted.  It
  is only in that context that effect sizes can become meaningful.  If a
 
 Even then standardized effect sizes may not be very helpful. We need
 to know much more information about the effect, the sensitivity of
 our measurements and so on.
 
 Thom
 
 
 =
 Instructions for joining and leaving this list and remarks about
 the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
 =




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-14 Thread Rich Ulrich

I think, some other folks are being sloppy about effect sizes.

Power analysis for the social sciences  is a book that
defines small, medium and large effects in terms that are
convenient and  *usually*  appropriate 
for the *social sciences* -- it makes no pretenses 
that these are universally applicable.

Similarly -- 

On Thu, 13 Sep 2001 18:17:54 -0500, jim clark [EMAIL PROTECTED]
wrote:

 Hi
 
 I found the Rosenthal reference that addresses the following
 point:
 
 On 13 Sep 2001, Herman Rubin wrote:
  The effect size is NOT small, or it would not save more
  than a very small number of lives.  If it were small,
  considering the dangers of aspirin, it would not be used
  for this purpose.
 
 At http://davidmlane.com/hyperstat/B165665.html, one finds:
 
 Rosenthal (1990) showed that although aspirin cut the risk of a
 heart attack approximately in half, it explained only .0011 of
 the variance (.11%). Similarly, Abelson (1985) found that batting
 average accounted for a trivial proportion of the variance in
 baseball game outcomes.  Therefore, measures of proportion of
 variance explained do not always communicate the importance of an
 effect accurately.

Stripping down the verbiage, that says: 
[M]easures of ... variance ... do not ... communicate ... the
effect.

That is exactly what Herman asserted.  A two-fold effect is 
moderate in epidemiology, whether you need 500 or 10,000
subjects-per-group  to detect it.   A five-fold effect is large.
In epi, these are measured by the odds ratio (which is also 
the Relative risk when the occurrence rates are low).

Epidemiologists do not use R-squared as an effect measure 
because it varies 20-fold for the same 500 versus 10,000.
R^2  is robust as a *test*  but it is not robust  where the 
odds ratio is robust, as a measure of effect size.
 
 The reference for Rosenthal is:
 
 Rosenthal, R. (1990). How are we doing in soft psychology?
 American Psychologist, 45, 775-777.
-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Rolf Dalin

Hi, this is about Jim Clark's reply to dennis roberts.

 On 12 Sep 2001, dennis roberts wrote:
  At 07:23 PM 9/12/01 -0500, jim clark wrote:
  What your table shows is that _both_ dimensions are informative.
  That is, you cannot derive effect size from significance, nor
  significance from effect size.  To illustrate why you need both,
  consider a study with small n that happened to get a large effect
  that was not significant.  The large effect should be ignored
  as being due to chance.  Only having the effect size would likely
  lead to the error of treating it as real (i.e., non-chance.
  
  or, another way to view it is that neither of the dimensions is very
  informative
 
 I'm not sure how both informative gets translated into neither
 very informative.  Seems like a perverse way of thinking to me.  
 Moreover, your original question was then what benefit is there
 to look at significance AT ALL? which implied to me that your
 view was that significance was not important and that effect
 size conveyed all that was needed.

When using the information conveyed in the p-values and/or effect 
size measures and/or decisions about some null hypothesis, in my 
opinion there's only one place to look: effect size measures given 
with CIs are informative. Significance alone gives you no clue to 
whether an effect is of any practical importance in the real world 
situation.

...

  the distinction between significant or not ... is based on an arbitrary
  cutoff point ... which has on one side ... the notion that the null
  seems as though it might be tenable ... and the other side ... the
  notion that the null does not seem to tenable ... but this is not an
  either/or deal ... it is only a matter of degree
 
 It was your table, but the debate would be the same if you put
 multiple rows with p values along the dimension.  That is, what
 is the relative importance of significance (or p value) and
 effect size.
 
Yes it would be the same debate. No matter how small the p-value it 
gives very little information about the effect size or its practical 
importance. 

When your data are on a scale which is arbitrary, not meters or 
USDs, let's say you have constructed a scale from multiple items. 
How do you define effect size? How can differences between means 
be interpreted to be informative? 

Cheers! /Rolf Dalin
**
Rolf Dalin
Department of Information Tchnology and Media
Mid Sweden University
S-870 51 SUNDSVALL
Sweden
Phone: 060 148690, international: +46 60 148690
Fax: 060 148970, international: +46 60 148970
Mobile: 0705 947896, intnational: +46 70 5947896
Home: 060 21933, intnational: +46 60 21933
mailto:[EMAIL PROTECTED]
http://www.itk.mh.se/~roldal/
**


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Thom Baguley

Rolf Dalin wrote:
 Yes it would be the same debate. No matter how small the p-value it
 gives very little information about the effect size or its practical
 importance.

Neither do standardized effect sizes.

Thom


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread jim clark

Hi

On 13 Sep 2001, Rolf Dalin wrote:
 Hi, this is about Jim Clark's reply to dennis roberts.
  I'm not sure how both informative gets translated into neither
  very informative.  Seems like a perverse way of thinking to me.  
  Moreover, your original question was then what benefit is there
  to look at significance AT ALL? which implied to me that your
  view was that significance was not important and that effect
  size conveyed all that was needed.
 
 When using the information conveyed in the p-values and/or effect 
 size measures and/or decisions about some null hypothesis, in my 
 opinion there's only one place to look: effect size measures given 
 with CIs are informative. Significance alone gives you no clue to 
 whether an effect is of any practical importance in the real world 
 situation.

I did not say anything about significance alone and (I
thought) clearly indicated that it was important to consider such
things as effect size, although effect sizes are no less
difficult to use intelligently than significance tests (or
confidence intervals).

   the distinction between significant or not ... is based on an arbitrary
   cutoff point ... which has on one side ... the notion that the null
   seems as though it might be tenable ... and the other side ... the
   notion that the null does not seem to tenable ... but this is not an
   either/or deal ... it is only a matter of degree
  
  It was your table, but the debate would be the same if you put
  multiple rows with p values along the dimension.  That is, what
  is the relative importance of significance (or p value) and
  effect size.
  
 Yes it would be the same debate. No matter how small the p-value it 
 gives very little information about the effect size or its practical 
 importance. 

Sorry I have to disagree.  If you do not have a reasonably small
significance level (or equivalently, a CI that does not show some
evidence of non-overlap with a null effect), then any effort to
study effect size or consider practical significance is wishful
thinking.

 When your data are on a scale which is arbitrary, not meters or 
 USDs, let's say you have constructed a scale from multiple items. 
 How do you define effect size? How can differences between means 
 be interpreted to be informative? 

To my knowledge effect size does not depend on the kind of
measure.  Many clinical scales in psychology, for example, are
composite measures and effect sizes and practical significance
can be computed and readily interpreted, albeit with some caution
(not due to the nature of the measure but to the subtleties of
all statistal methods).  A clinician, for example, can indicate
how much of a reduction on a measure of clinical depression would
be clinically important.

Sometimes I think that people are looking for some magic
bullet in statistics (i.e., significance, effect size,
whatever) that is going to avoid all of the problems and
misinterpretations that arise from existing practices. I think
that is a naive belief and that we need to teach how to use all
of the tools wisely because all of the tools are prone to abuse
and misinterpretation.

Best wishes
Jim


James M. Clark  (204) 786-9757
Department of Psychology(204) 774-4134 Fax
University of Winnipeg  4L05D
Winnipeg, Manitoba  R3B 2E9 [EMAIL PROTECTED]
CANADA  http://www.uwinnipeg.ca/~clark




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Dennis Roberts

At 02:33 PM 9/13/01 +0100, Thom Baguley wrote:
Rolf Dalin wrote:
  Yes it would be the same debate. No matter how small the p-value it
  gives very little information about the effect size or its practical
  importance.

Neither do standardized effect sizes.

agreed ... of course, we would all be a whole lot better off if a 
researcher DEFINED for us ...  a priori ... what size of effect he/she 
considered to be important ... and why that amount has practical benefit

then we could evaluate (approximately) if he/she found something of 
practical importance (at least according to the researcher)

but, what we get is an after the fact description of this  which by the 
very nature of its post hocness ... is not really that helpful

bringing into this discussion statistical significance only has relevance 
IF you believe that null hypothesis testing is of real value

bringing  effect size into the discussion only has relevance IF you know 
what effect it takes to have practical importance ... (and i have yet to 
see the article that focuses on this even if they do report effect sizes ... )

what we need in all of this is REPLICATION ... and, the accumulation of 
evidence about the impact of independent variables that we consider to have 
important potential ... and not to waste our time and money on so many 
piddly manipulations ... just for the sake of getting stuff published

the recent push by apa journals and others to make submitters supply 
information on effect sizes is, in my view, misplaced effort ... what 
should be insisted upon in studies where the impacts of variables is being 
investigated ... is a clear statement and rationale BY the researcher as to 
WHAT size impact it would take to make a practical and important difference 
...

if that ONE piece of information were insisted upon ... then all of us 
would be in a much better position to evaluate results that are presented

reporting effect sizes (greater than 0) does nothing to help readers 
understand the functional benefit that might result from such treatments

until we tackle that issue directly, we are more or less going around in 
circles


Thom


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: effect size/significance

2001-09-13 Thread Donald Burrill

On Thu, 13 Sep 2001, Paul R. Swank wrote in part:

 Dennis said
 
 other than being able to say that the experimental group ... ON AVERAGE ...
 had a mean that was about 1.11 times (control group sd units) larger than
 the control group mean, which is purely DESCRIPTIVE ... what  can you say
 that is important?
 
 However, can you say even that unless it is ratio scale?

Yes, well, Dennis was referring to a _difference_.  When the underlying 
scale is interval, differences ARE ratio scale:  zero means zero.
-- Don.
 
 Donald F. Burrill [EMAIL PROTECTED]
 184 Nashua Road, Bedford, NH 03110  603-471-7128



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Alan McLean

jim clark wrote:
 

 
 Sometimes I think that people are looking for some magic
 bullet in statistics (i.e., significance, effect size,
 whatever) that is going to avoid all of the problems and
 misinterpretations that arise from existing practices. I think
 that is a naive belief and that we need to teach how to use all
 of the tools wisely because all of the tools are prone to abuse
 and misinterpretation.
 


Spot on!
Alan

-- 
Alan McLean ([EMAIL PROTECTED])
Department of Econometrics and Business Statistics
Monash University, Caulfield Campus, Melbourne
Tel:  +61 03 9903 2102Fax: +61 03 9903 2007


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread jim clark

Hi

On 13 Sep 2001, Herman Rubin wrote:
 jim clark  [EMAIL PROTECTED] wrote:
 Or consider a study with a small effect size that is significant.  
 The fact that the effect is significant indicates that some
 non-chance effect is present and it might very well be important
 theoretically or even practically despite the small effect size.  
 The classic example of practical significance is the use of
 aspirin prophylactically to prevent cardio-vascular problems.  
 The effect size is quite small, but the effect was significant in
 clinical studies and, given the large numbers of people involved,
 was of great practical importance (i.e., many lives were
 extended).
 
 The effect size is NOT small, or it would not save more
 than a very small number of lives.  If it were small,
 considering the dangers of aspirin, it would not be used
 for this purpose.

I was surprised by Herman's response here.  I was going by
memory, but I thought that someone (Rosenthal perhaps?) had
demonstrated that conventional effect size statistics for the
effect of aspirin were indeed small.  The great benefit comes
from a small effect that is applied to many, many people.  An
effect of .01 applied to 1,000,000 people represents 10,000
lives.  I'll try to track it down.

Best wishes
Jim


James M. Clark  (204) 786-9757
Department of Psychology(204) 774-4134 Fax
University of Winnipeg  4L05D
Winnipeg, Manitoba  R3B 2E9 [EMAIL PROTECTED]
CANADA  http://www.uwinnipeg.ca/~clark




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread jim clark

Hi

I found the Rosenthal reference that addresses the following
point:

On 13 Sep 2001, Herman Rubin wrote:
 The effect size is NOT small, or it would not save more
 than a very small number of lives.  If it were small,
 considering the dangers of aspirin, it would not be used
 for this purpose.

At http://davidmlane.com/hyperstat/B165665.html, one finds:

Rosenthal (1990) showed that although aspirin cut the risk of a
heart attack approximately in half, it explained only .0011 of
the variance (.11%). Similarly, Abelson (1985) found that batting
average accounted for a trivial proportion of the variance in
baseball game outcomes.  Therefore, measures of proportion of
variance explained do not always communicate the importance of an
effect accurately.

The reference for Rosenthal is:

Rosenthal, R. (1990). How are we doing in soft psychology?
American Psychologist, 45, 775-777.

Best wishes
Jim


James M. Clark  (204) 786-9757
Department of Psychology(204) 774-4134 Fax
University of Winnipeg  4L05D
Winnipeg, Manitoba  R3B 2E9 [EMAIL PROTECTED]
CANADA  http://www.uwinnipeg.ca/~clark




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Dennis Roberts

here are some data ... say we randomly assigned 30 Ss ... 15 to each 
condition and found the following:

MTB  desc c1 c2

Descriptive Statistics: exp, cont


Variable N   Mean Median TrMean  StDevSE Mean
exp 15  26.13  27.00  26.00   4.96   1.28
cont15  21.73  22.00  21.85   3.95   1.02

MTB  twos c1 c2

Two-sample T for exp vs cont

N  Mean StDev   SE Mean
exp   15 26.13  4.96   1.3
cont  15 21.73  3.95   1.0

Difference = mu exp - mu cont
Estimate for difference:  4.40
95% CI for difference: (1.04, 7.76)
T-Test of difference = 0 (vs not =): T-Value = 2.69  P-Value = 0.012  
p value

MTB  let k1=(26.13-21.73)/3.95
MTB  prin k1

Data Display

K11.11392   simple effect size calculation

other than being able to say that the experimental group ... ON AVERAGE ... 
had a mean that was about 1.11 times (control group sd units) larger than 
the control group mean, which is purely DESCRIPTIVE ... what  can you say 
that is important?




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Elliot Cramer

Dennis Roberts [EMAIL PROTECTED] wrote:
: given a simple effect size calculation ... some mean difference compared to 

: that is ... can we not get both NS or sig results ... when calculated 
: effect sizes are small, medium, or large?

: if that is true ... then what benefit is there to look at significance AT ALL

effect size does not depend on the standard error of the estimate so one
can get a large nonsignificant effect.  Obviously you want some reason to
believe that the large estimated effect is real.


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Mike Granaas

On Thu, 13 Sep 2001, Dennis Roberts wrote:
 see the article that focuses on this even if they do report effect sizes ... )
 
 what we need in all of this is REPLICATION ... and, the accumulation of 
 evidence about the impact of independent variables that we consider to have 
 important potential ... and not to waste our time and money on so many 
 piddly manipulations ... just for the sake of getting stuff published

I was wondering when replication was going to get brought into this
discussion.  It's in all the textbooks, but we seem to forget it when it
comes time to publish...can replications without mutation even be
published?  

At the end of the day the individual study gives us data that is
consistent, inconsistent, or ambigious about something.  We add that bit
of information to the balance scales until one of the arms goes down and
stays down...then we truly know something.

 
 the recent push by apa journals and others to make submitters supply 
 information on effect sizes is, in my view, misplaced effort ... what 
 should be insisted upon in studies where the impacts of variables is being 
 investigated ... is a clear statement and rationale BY the researcher as to 
 WHAT size impact it would take to make a practical and important difference 
 ...

I think that we might agree:  I would say that studies need a clear a
priori rational (theoretical or empirical) prior to being conducted.  It
is only in that context that effect sizes can become meaningful.  If a
study was done just 'cause then we will frequently not be able to make
sense of the effect size measures.

Michael
 
 if that ONE piece of information were insisted upon ... then all of us 
 would be in a much better position to evaluate results that are presented
 
 reporting effect sizes (greater than 0) does nothing to help readers 
 understand the functional benefit that might result from such treatments
 
 until we tackle that issue directly, we are more or less going around in 
 circles
 
 

***
Michael M. Granaas
Associate Professor[EMAIL PROTECTED]
Department of Psychology
University of South Dakota Phone: (605) 677-5295
Vermillion, SD  57069  FAX:   (605) 677-6604
***
All views expressed are those of the author and do not necessarily
reflect those of the University of South Dakota, or the South
Dakota Board of Regents.



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: effect size/significance

2001-09-13 Thread Paul R. Swank

Dennis said

other than being able to say that the experimental group ... ON AVERAGE ...
had a mean that was about 1.11 times (control group sd units) larger than
the control group mean, which is purely DESCRIPTIVE ... what  can you say
that is important?

However, can you say even that unless it is ratio scale?
OTOH, there is a two standard deviation difference, which is large enough to
practically hit you over the head. The significance test is more important
when the effect is smaller because small effects are much more likely to be
chance events.

Paul R. Swank, Ph.D.
Professor
Developmental Pediatrics
UT Houston Health Science Center

=




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Herman Rubin

In article [EMAIL PROTECTED],
jim clark  [EMAIL PROTECTED] wrote:
Hi

On 13 Sep 2001, Rolf Dalin wrote:
 Hi, this is about Jim Clark's reply to dennis roberts.

.

Sometimes I think that people are looking for some magic
bullet in statistics (i.e., significance, effect size,
whatever) that is going to avoid all of the problems and
misinterpretations that arise from existing practices. I think
that is a naive belief and that we need to teach how to use all
of the tools wisely because all of the tools are prone to abuse
and misinterpretation.

This is what all those who have only taken methods courses,
or in fact not studied it from a decision approach CAN do.
The discussion of significance, confidence intervals, etc.,
ignores some aspects of the value of the decision, or else
ignores some of the states of nature.  The decision approach
needs to be used with care, as the user's assumptions affect
the decision taken, and this cannot be ignored; in some cases,
the effects of some of the assumptions are small enough that
they are not of much importance, but few users know which.
-- 
This address is for information only.  I do not claim that these views
are those of the Statistics Department or of Purdue University.
Herman Rubin, Dept. of Statistics, Purdue Univ., West Lafayette IN47907-1399
[EMAIL PROTECTED] Phone: (765)494-6054   FAX: (765)494-0558


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-13 Thread Herman Rubin

In article [EMAIL PROTECTED],
jim clark  [EMAIL PROTECTED] wrote:
Hi

On 12 Sep 2001, Dennis Roberts wrote:

 that is ... can we not get both NS or sig results ... when calculated 
 effect sizes are small, medium, or large?

 if that is true ... then what benefit is there to look at
 significance AT ALL

...

What your table shows is that _both_ dimensions are informative.  
That is, you cannot derive effect size from significance, nor
significance from effect size.

This has to be drummed into the users of statistics.

 To illustrate why you need both,
consider a study with small n that happened to get a large effect
that was not significant.  The large effect should be ignored
as being due to chance.  Only having the effect size would likely
lead to the error of treating it as real (i.e., non-chance.

Or consider a study with a small effect size that is significant.  
The fact that the effect is significant indicates that some
non-chance effect is present and it might very well be important
theoretically or even practically despite the small effect size.  
The classic example of practical significance is the use of
aspirin prophylactically to prevent cardio-vascular problems.  
The effect size is quite small, but the effect was significant in
clinical studies and, given the large numbers of people involved,
was of great practical importance (i.e., many lives were
extended).

The effect size is NOT small, or it would not save more
than a very small number of lives.  If it were small,
considering the dangers of aspirin, it would not be used
for this purpose.

So we need to be sensitive to both the magnitude and significance
of our effects, and we need to be duly cautious in the
interpretation of both.

The way to take it all into account is to use the decision
approach.  This can be stated in one sentence:

It is necessary to consider all consequences of
the proposed action in all states of nature.

Consistency applied to this states that one should use
some weight measure on the utility evaluation of the
consequences for the various states of nature, and sum
or integrate.

-- 
This address is for information only.  I do not claim that these views
are those of the Statistics Department or of Purdue University.
Herman Rubin, Dept. of Statistics, Purdue Univ., West Lafayette IN47907-1399
[EMAIL PROTECTED] Phone: (765)494-6054   FAX: (765)494-0558


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-12 Thread Lise DeShea

At 04:04 PM 9/12/01 -0400, you wrote:

if that is true ... then what benefit is there
to look at significance AT ALL

To get published, get tenure, and avoid having to live in a cardboard box
in the park. Ha ha!

Lise


Re: effect size/significance

2001-09-12 Thread jim clark

Hi

On 12 Sep 2001, dennis roberts wrote:
 At 07:23 PM 9/12/01 -0500, jim clark wrote:
 What your table shows is that _both_ dimensions are informative.
 That is, you cannot derive effect size from significance, nor
 significance from effect size.  To illustrate why you need both,
 consider a study with small n that happened to get a large effect
 that was not significant.  The large effect should be ignored
 as being due to chance.  Only having the effect size would likely
 lead to the error of treating it as real (i.e., non-chance.
 
 or, another way to view it is that neither of the dimensions is very 
 informative

I'm not sure how both informative gets translated into neither
very informative.  Seems like a perverse way of thinking to me.  
Moreover, your original question was then what benefit is there
to look at significance AT ALL? which implied to me that your
view was that significance was not important and that effect
size conveyed all that was needed.

 of course we know that significance does not mean real and non 
 significance does not mean chance alone ... there is no way to decipher 
 one from the other based on our significance tests

I included in parentheses after the real (i.e., non-chance).  
That is exactly what significance means ... the probability of
this large a difference arising by chance (i.e., when Ho is
true) is small.  And non-significance does mean that the outcome
had a reasonably high probability (i.e., chance) of occurring if
the Ho were true.

 the distinction between significant or not ... is based on an arbitrary 
 cutoff point ... which has on one side ... the notion that the null seems 
 as though it might be tenable ... and the other side ... the notion that 
 the null does not seem to tenable ... but this is not an either/or deal ... 
 it is only a matter of degree

It was your table, but the debate would be the same if you put
multiple rows with p values along the dimension.  That is, what
is the relative importance of significance (or p value) and
effect size.

 what if we juxtapose ... non significant findings with a large effect size 
 ... with significant results with a small effect size ... which of these 
 two would most feel had most import?

Nonsignificant findings with a large effect size should be
ignored (except perhaps as some indication that it might be
worth replicating the study or gathering more observations to
increase power).  Significant results with a small effect size
may or may not be important (e.g., the aspirin example), but at
least there is little chance that they just occurred by chance.

 surely, if both dimensions are seen as been equal parts of the overall 
 puzzle, then, both would seem to be more or less equal

This is an over-simplification.  I would say that significance
has priority.  Effect sizes for non-significant effects are
meaningless except as described above.  Given significance, one
can turn attention to effect size, which itself is no simple
matter.

 but, if one opts for large effect size when results are not significant ... 
 then, this suggests that significance adds little if anything to the mix ...

One should never trust effect sizes that are not significant, no
matter how large.  I would be very interested in seeing a case
where this principle was violated to good effect.

 however, if we opt for significance along with a small effect
 size, then this suggests that significance is playing a more
 important role in one's eyes

Correct.  The issue of whether the small effect size is an issue
depends on numerous other factors.

 the reality is too that effect sizes are, when categorized as small, 
 medium, and large ... again ... totally arbitrary ... which makes it even 
 harder still to make much sense of these ... just like it is difficult to 
 make much sense out of significance

Are you thinking of retreating to some mountaintop and giving up
on the world of statistics?

 and finally, effect sizes do NOT say anything about importance OF the 
 effect ... merely some statement about the SIZE of the effect ... so, it 
 could very well be that for many independent variables ... a small effect 
 size has a much greater impact consequence than a large effect size ... 
 even when both are found to have significance on their side

That is correct, as in the example I gave.

 just to muddy the waters more

And here I thought the waters were muddy all ready!

Best wishes
Jim


James M. Clark  (204) 786-9757
Department of Psychology(204) 774-4134 Fax
University of Winnipeg  4L05D
Winnipeg, Manitoba  R3B 2E9 [EMAIL PROTECTED]
CANADA  http://www.uwinnipeg.ca/~clark




=
Instructions for joining and leaving 

Re: effect size/significance

2001-09-12 Thread jim clark

Hi

On 12 Sep 2001, Dennis Roberts wrote:
 given a simple effect size calculation ... some mean difference compared to 
 some pooled group or group standard deviation ... is it not possible to 
 obtain the following combinations (assuming some significance test is done)
 
effect size
   small   mediumlarge
 
  res NS
 
  res sig
 
 that is ... can we not get both NS or sig results ... when calculated 
 effect sizes are small, medium, or large?
 
 if that is true ... then what benefit is there to look at
 significance AT ALL

What your table shows is that _both_ dimensions are informative.  
That is, you cannot derive effect size from significance, nor
significance from effect size.  To illustrate why you need both,
consider a study with small n that happened to get a large effect
that was not significant.  The large effect should be ignored
as being due to chance.  Only having the effect size would likely
lead to the error of treating it as real (i.e., non-chance.

Or consider a study with a small effect size that is significant.  
The fact that the effect is significant indicates that some
non-chance effect is present and it might very well be important
theoretically or even practically despite the small effect size.  
The classic example of practical significance is the use of
aspirin prophylactically to prevent cardio-vascular problems.  
The effect size is quite small, but the effect was significant in
clinical studies and, given the large numbers of people involved,
was of great practical importance (i.e., many lives were
extended).

So we need to be sensitive to both the magnitude and significance
of our effects, and we need to be duly cautious in the
interpretation of both.

Best wishes
Jim


James M. Clark  (204) 786-9757
Department of Psychology(204) 774-4134 Fax
University of Winnipeg  4L05D
Winnipeg, Manitoba  R3B 2E9 [EMAIL PROTECTED]
CANADA  http://www.uwinnipeg.ca/~clark




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: effect size/significance

2001-09-12 Thread dennis roberts

At 07:23 PM 9/12/01 -0500, jim clark wrote:
Hi


What your table shows is that _both_ dimensions are informative.
That is, you cannot derive effect size from significance, nor
significance from effect size.  To illustrate why you need both,
consider a study with small n that happened to get a large effect
that was not significant.  The large effect should be ignored
as being due to chance.  Only having the effect size would likely
lead to the error of treating it as real (i.e., non-chance.



or, another way to view it is that neither of the dimensions is very 
informative

of course we know that significance does not mean real and non 
significance does not mean chance alone ... there is no way to decipher 
one from the other based on our significance tests

the distinction between significant or not ... is based on an arbitrary 
cutoff point ... which has on one side ... the notion that the null seems 
as though it might be tenable ... and the other side ... the notion that 
the null does not seem to tenable ... but this is not an either/or deal ... 
it is only a matter of degree

what if we juxtapose ... non significant findings with a large effect size 
... with significant results with a small effect size ... which of these 
two would most feel had most import?

surely, if both dimensions are seen as been equal parts of the overall 
puzzle, then, both would seem to be more or less equal

but, if one opts for large effect size when results are not significant ... 
then, this suggests that significance adds little if anything to the mix ...

however, if we opt for significance along with a small effect size, then 
this suggests that significance is playing a more important role in one's eyes

the reality is too that effect sizes are, when categorized as small, 
medium, and large ... again ... totally arbitrary ... which makes it even 
harder still to make much sense of these ... just like it is difficult to 
make much sense out of significance

and finally, effect sizes do NOT say anything about importance OF the 
effect ... merely some statement about the SIZE of the effect ... so, it 
could very well be that for many independent variables ... a small effect 
size has a much greater impact consequence than a large effect size ... 
even when both are found to have significance on their side

just to muddy the waters more


==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=