Re: SRSes

2001-07-25 Thread Art Kendall

my previous remarks were about other sampling designs.  I was comaring valid
complex designs to SRS design and not non-sampling case selection.

dennis roberts wrote:

> my hypothesis of course is that more often than not ... in data collection
> problems where sampling is involved AND inferences are desired ... we goof
> far more often ... than do a better than SRS job of sampling
>
> 1. i wonder if anyone has really taken a SRS of the literature ... maybe
> stratified by journals or disciplines ... and tried to see to what extent
> sampling in the investigations was done via SRS ... better than that ... or
> worse than that??? of course, i would expect even if this is done ... we
> would have a + biased figure ... since, the notion is that only the
> better/best of the submitted stuff gets published so, the figures for all
> stuff that is done (ie, the day in day out batch), published or not ...
> would have to look worse off ...
>
> 2. can worse than SRS ... be as MUCH worse ... as complex sampling plans
> can be better than SRS??? that is ... could a standard error for a bad
> sampling plan (if we could even estimate it) ... be proportionately as much
> LARGER than the standard error for SRS samples ... as complex sampling
> plans can produce standard errors that are as proportionately SMALLER than
> SRS samples? are there ANY data that exist on this matter?
>
> ==
> dennis roberts, penn state university
> educational psychology, 8148632401
> http://roberts.ed.psu.edu/users/droberts/drober~1.htm
>
> =
> Instructions for joining and leaving this list and remarks about
> the problem of INAPPROPRIATE MESSAGES are available at
>   http://jse.stat.ncsu.edu/
> =



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: SRSes

2001-07-25 Thread Art Kendall



Dennis Roberts wrote:

> 
>
> but, we KNOW that most samples are drawn in a way that is WORSE than SRS ...
>
> thus, essentially every CI ... is too narrow ... or, every test statistic
> ... t or F or whatever ... has a p value that is too LOW ...
>
> what adjustment do we make for this basic problem?

The adjustment for design is done with weights to get the point estimates using
regular software such as SPSS etc.  To get the confidence estimates special
software such as WESVAR, SUDAAN, or CPLEX is commonly used.  Because the latter
packages are not as user friendly in their presentation of results, I usually
get the point estimates in SPSS, then I use WESVAR or SUDAAN and get both point
and interval estimates.  I use the point estimates from the latter packages as
"navigational aids" to find the interval estimates in the output and to assure
that I am getting the right computations.

Some sampling designs include cluster sampling (random effects), some
stratification (fixed effects), and some both.
For those with stratification only, if there is any difference in the means
(proportions) among the strata, usually the CIs will be too wide.  For those
with cluster sampling, usually the CIs will be too narrow.  For those designs
with both stratification and clustering, the CIs will be subject to both
narrowing and widening, and only specilized software will tell the net effect.

In addition, ratio, regression, or difference estimates may have narrower "true"
CIs.



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: SRSes

2001-07-24 Thread dennis roberts

my hypothesis of course is that more often than not ... in data collection 
problems where sampling is involved AND inferences are desired ... we goof 
far more often ... than do a better than SRS job of sampling

1. i wonder if anyone has really taken a SRS of the literature ... maybe 
stratified by journals or disciplines ... and tried to see to what extent 
sampling in the investigations was done via SRS ... better than that ... or 
worse than that??? of course, i would expect even if this is done ... we 
would have a + biased figure ... since, the notion is that only the 
better/best of the submitted stuff gets published so, the figures for all 
stuff that is done (ie, the day in day out batch), published or not ... 
would have to look worse off ...

2. can worse than SRS ... be as MUCH worse ... as complex sampling plans 
can be better than SRS??? that is ... could a standard error for a bad 
sampling plan (if we could even estimate it) ... be proportionately as much 
LARGER than the standard error for SRS samples ... as complex sampling 
plans can produce standard errors that are as proportionately SMALLER than 
SRS samples? are there ANY data that exist on this matter?


==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: SRSes

2001-07-24 Thread Simon, Steve, PhD

Dennis Roberts writes:

> most books talk about inferential statistics ... particularly those 
> where you take a sample ... find some statistic ... estimate some error 
> term ... then build a CI or test some null hypothesis ...
> 
> error in these cases is always assumed to be based on taking AT LEAST a 
> simple random sample ... or SRS as some books like to say ...
> 
> but, we KNOW that most samples are drawn in a way that is WORSE than SRS 

> thus, essentially every CI ... is too narrow ... or, every test 
> statistic ... t or F or whatever ... has a p value that is too LOW  
> 
> what adjustment do we make for this basic problem?

Another thought provoking question from Penn State.

In the real world, most people assess the deviation from SRS in a
qualitative (non-quantitative) fashion. If the deviation is serious, then
you consider it as a more preliminary finding or one that is in greater need
of replication. If it is very serious, you totally disregard the findings
from the study. The folks in Evidence Based Medicine talk about levels of
evidence, and this is one of the things that they would use to select
whether a study represents a higher or lower level of evidence.

You probably do the same thing when you assess problems with non-response
bias, recall bias, and subjects who drop out in the middle of the study.
Typically you assess these in a qualitative fashion because it is so
difficult to quantify how much these will bias your findings.

You could argue that this represents the classic distinction between
sampling error and non-sampling error. The classic CI is almost always too
narrow, because it only accounts for some of the uncertainty in the model.
We are getting more sophisticated, but we still can't quantify many of the
additional sources of uncertainty.

By the way, if you take non-SRS sample and then randomly allocate these
patients to a treatment and control group, the CI appropriately accounts for
uncertainty within this population, but you have trouble extrapolating to
the population that you are more interested in. It's the classic internal
versus external validity argument.

I hope this makes sense and is helpful.

Steve Simon, [EMAIL PROTECTED], Standard Disclaimer.
STATS: STeve's Attempt to Teach Statistics. http://www.cmh.edu/stats
Watch for a change in servers. On or around June 2001, this page will
move to http://www.childrens-mercy.org/stats



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: SRSes

2001-07-24 Thread Jerry Dallal

Dennis Roberts wrote:

> but, we KNOW that most samples are drawn in a way that is WORSE than SRS ...
> 
> thus, essentially every CI ... is too narrow ... or, every test statistic
> ... t or F or whatever ... has a p value that is too LOW ...
> 
> what adjustment do we make for this basic problem?

We do it anyway!  The real concern isn't that CIs are to narrow or
that Ps are too liberal, but that they are completely irrelevant. 
If it's impossible to specify a sampling model, there's no formal
basis for inference.  (I'm ignoring randomized trials, which can be
valid without being generalizable.)  

For better or worse, here's what I tell my students in an attempt at
honesty...

Sometimes the pedigree of a sample is uncertain, yet standard
statistical techniques for simple random samples are used. The
rationale behind such analyses is best expressed in a reworking of a
quotation from Stephen Fienberg, in which the phrases (contingency
table and multinomial have been replaced by survey and simple
random): 

 "It is often true that data in a [survey] have not been
produced by a [simple random] sampling procedure, and that the
statistician is unable to determine the exact sampling scheme which
was used. In such situations the best we can do, usually, is to
assume a [simple random] situation and hope that it is not very
unreasonable."

This does not mean that sampling issues can be disregarded. Rather,
it says that in some instances we may treat data as though they
arose from a simple random sample, barring evidence that such an
approach is inappropriate.


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: SRSes

2001-07-24 Thread Dennis Roberts

At 03:55 PM 7/24/01 -0400, Donald Burrill wrote:
>Hi, Dennis!
> Yes, as you point out, most elementary textbooks treat only SRS
>types of samples.  But while (as you also point out) some more realistic
>sampling methods entail larger sampling variance than SRS, some of them
>have _smaller_ variance -- notably, stratified designs when the strata
>differ between themselves on the quantity being measured.

sure ... i know that

(then i said) ... but, we KNOW that most samples are drawn in a way that is 
WORSE than SRS



and you responded

>I don't think _I_ know this.  I know that SOME samples are so drawn;
>but (see above) I also know that SOME samples are drawn in a way that
>is BETTER than SRS (where I assume by "worse" you meant "with larger
>sampling variance", so by "better" I mean "with smaller sampling
>variance").

i think we do know this ... if you enumerate all the situations you know of 
where sampling from some larger population has been done ... i would bet a 
dollar to a penny that ... the sampling plan is WORSE than SRS  ... so, i 
would suggest that the NORM is worse ... the exception is SRS or better

i don't think books spend nearly enough time ... on the fact that most day 
in day out samples are taken in a pretty pathetic way ...


>I perceive the "basic problem" as the fact that sampling variance is
>(relatively) easily calculated for a SRS, while it is more difficult
>to calculate under almost _any_ other type of sampling.

sure ... but, books ONLY seem to discuss the easy way ... and i do too ... 
because it seems rather straight forward ... but, given time constraints 
... it never goes further than that ...

>  Whether it is enough more difficult that one would REALLY like to avoid
>it in an elementary course is a judgement call;  but for the less
>quantitatively-oriented students with whom many of us have to deal, we
>_would_ often like to avoid those complications.  Certainly dealing with
>the completely _general_ case is beyond the scope of a first course, so
>it's just a matter of deciding how many, and which, specific types of
>cases one is willing to shoehorn into the semester (and what "previews
>of coming attractions" one wishes to allude to in higher-level courses).

however, we do become sticklers for details ... and force students to use 
the correct CVs, make the right CIs, ... do the t tests correctly ... and 
heaven forbib if you get off a line or two when reading off the values from 
the t table ...


>Seems to me the most sensible "adjustment" (and of a type we make at
>least implicitly in a lot of other areas too) is
>  = to acknowledge that the calculations for SRS are presented
>(a) for a somewhat unrealistic "ideal" kind of case,

i would stress ... really unrealistic ...

>(b) to give the neophyte _some_ experience in playing this game,

and then leave them hanging

>Some textbooks I have used (cf. Moore, "Statistics:  Concepts &
>Controversies" (4th ed.), Table 1.1, page 40) present a table giving the
>margin of error for the Gallup poll sampling procedure, as a function of
>population percentage and sample size.  Such a table permits one to show
>how Gallup's precision varies from what one would calculate for a SRS,
>thus providing some small emphasis for the cautionary tale one wishes to
>convey.

but ... in moore and mccabe ... the stress throughout the book ... is on 
SRSes ... and no real mention is made nor solutions to ... the problems 
that it will be a rare day in analysis land ... for the typical person 
working with data ... to be doing SRS sampling ...
it's just not going to happen

the bottom line, IMHO, is that we glide over this like it is not a problem 
at all ... when we know it is


>  
>  Donald F. Burrill [EMAIL PROTECTED]
>  184 Nashua Road, Bedford, NH 03110  603-471-7128

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: SRSes

2001-07-24 Thread Donald Burrill

Hi, Dennis!
Yes, as you point out, most elementary textbooks treat only SRS 
types of samples.  But while (as you also point out) some more realistic 
sampling methods entail larger sampling variance than SRS, some of them 
have _smaller_ variance -- notably, stratified designs when the strata 
differ between themselves on the quantity being measured.

On Tue, 24 Jul 2001, Dennis Roberts wrote:

> most books talk about inferential statistics ... particularly those 
> where you take a sample ... find some statistic ... estimate some error 
> term ... then build a CI or test some null hypothesis ...
> 
> error in these cases is always assumed to be based on taking AT LEAST a 
> simple random sample ... or SRS as some books like to say ...
> 
> but, we KNOW that most samples are drawn in a way that is WORSE than SRS 

I don't think _I_ know this.  I know that SOME samples are so drawn;  
but (see above) I also know that SOME samples are drawn in a way that 
is BETTER than SRS (where I assume by "worse" you meant "with larger 
sampling variance", so by "better" I mean "with smaller sampling 
variance").

> thus, essentially every CI ... is too narrow ... or, every test 
> statistic ... t or F or whatever ... has a p value that is too LOW  
> 
> what adjustment do we make for this basic problem?

I perceive the "basic problem" as the fact that sampling variance is 
(relatively) easily calculated for a SRS, while it is more difficult 
to calculate under almost _any_ other type of sampling.  
 Whether it is enough more difficult that one would REALLY like to avoid 
it in an elementary course is a judgement call;  but for the less 
quantitatively-oriented students with whom many of us have to deal, we 
_would_ often like to avoid those complications.  Certainly dealing with 
the completely _general_ case is beyond the scope of a first course, so 
it's just a matter of deciding how many, and which, specific types of 
cases one is willing to shoehorn into the semester (and what "previews 
of coming attractions" one wishes to allude to in higher-level courses). 

Seems to me the most sensible "adjustment" (and of a type we make at 
least implicitly in a lot of other areas too) is 
 = to acknowledge that the calculations for SRS are presented 
   (a) for a somewhat unrealistic "ideal" kind of case,
   (b) to give the neophyte _some_ experience in playing this game,
   (c) to see how the variance depends (apart from the sampling scheme)
on the sample size (and on the estimated value, if one is 
estimating proportions or percentages),
   (d) in despite of the fact that most real sampling is carried out 
under distinctly non-SRS conditions, and therefore entails 
variances for which SRS calculations may be quite awry;  and
 = to have yet another situation for which one can point out that for 
actually DOING anything like this one should first consult a 
competent statistician (or, perhaps, _become_ one!).

Some textbooks I have used (cf. Moore, "Statistics:  Concepts & 
Controversies" (4th ed.), Table 1.1, page 40) present a table giving the 
margin of error for the Gallup poll sampling procedure, as a function of 
population percentage and sample size.  Such a table permits one to show 
how Gallup's precision varies from what one would calculate for a SRS, 
thus providing some small emphasis for the cautionary tale one wishes to 
convey.

 
 Donald F. Burrill [EMAIL PROTECTED]
 184 Nashua Road, Bedford, NH 03110  603-471-7128



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



SRSes

2001-07-24 Thread Dennis Roberts

most books talk about inferential statistics ... particularly those where 
you take a sample ... find some statistic ... estimate some error term ... 
then build a CI or test some null hypothesis ...

error in these cases is always assumed to be based on taking AT LEAST a 
simple random sample ... or SRS as some books like to say ...

but, we KNOW that most samples are drawn in a way that is WORSE than SRS ...

thus, essentially every CI ... is too narrow ... or, every test statistic 
... t or F or whatever ... has a p value that is too LOW ...

what adjustment do we make for this basic problem?

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=