Omissions in Journal Articles

2001-05-03 Thread Neil W. Henry

There was a recent discussion here of errors in journal articles. A
related topic is incomplete information, or at least what I consider
incomplete information.

A recent article in the American Journal of Epidemiology (2001, Vol 153,
No 6, 596-603) contains some nicely laid out (and badly titled, but save
that for another thread) tables that show the mean value of a dependent
variable in male and female subsamples, broken down by a dichotomous
independent variable (exercise? Yes/No), and "adjusted"  first by age
and then by age and several other numerical variables.

In a sociological journal I would most often see this analysis reported
in multiple regression form, but again, that's another thread. In health
related journals the convention is to speak of adjusted means, i.e. the
predicted dependent variable value for members of each category of the
independent variable, with the other predictors ("covariates") set to
specific values.

The article does not specify what values of age (and the other
predictors) were used to create the adjusted means. Instead there is a
footnote to the table that says:
"Adjusted means calculated by using analysis of covariance."

My question, directed to those of you who are more familiar with
journals in this area than I am, is whether this is a standard footnote
/ explanation, which is supposed to make clear to regular readers what
has been done?

>From my perspective it is inadequate, since the ANCOVA (or regression
analysis) has merely produced a predictive formula, and any values
whatsoever of the covariates could be plugged in to the equation.

Now I happen to know what SPSS v10 does when asked to produce
"estimated" means in its univariate GLM procedure: it plugs in the mean
values. The output actually contains the values of those means for the
record. (Is this true with other statistical software?) A user who knows
the syntax can actually specify the values, but the Windows point+click
screen doesn't allow that. Here's an example of the default subcommand
statement:
/EMMEANS = TABLES(exercise) WITH(age=MEAN xother=MEAN)

Using the mean can produce misleading adjusted values, especially when
the table contains subsample comparisons as in this article, where "all
analyses were sex-specific". If the default SPSS ANCOVA were followed,
the adjusted values would be created at different values of age, since
mean age differs by sex, and other covariate means are also different in
this study. (In addition, different tables contained analyses of 2 other
subgroups of the sample, with different mean Xs.)

This may or may not be a problem for the authors' interpretation of the
results, but it seems reasonable to expect editors to be more sensitive
to their readers' need to know exactly what is going on.

--
  **
 `o^o' * Neil W. Henry ([EMAIL PROTECTED]) *
 -<:>- * Virginia Commonwealth University   *
 _/ \_ * Richmond VA 23284-2014  **
  *   http://www.people.vcu.edu/~nhenry   *
  ***




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Random Sampling and External Validity

2001-03-26 Thread Neil W. Henry


--9F3BD71D2EDA80683B2FF9EA
Content-Type: text/plain; charset=us-ascii
Content-Transfer-Encoding: 7bit


dennis roberts wrote:

> At 12:56 PM 3/25/01 -0500, Karl L. Wuensch wrote:

> >Here is how I resolve that problem:  Define the population from the sample,
> >rather than vice versa -- that is, my results can be generalized to any
> >population for which my sample could be reasonably considered to be a random
> >sample.  Maybe we could call this "transcendental sampling" ;-) -- it is
> >somewhat like transcendental realism, defining reality from our percetion of
> >it, eh?
>
> this sounds like  the method of grounded theory in the qualitative bailiwick
> ...
> look at data you have and see what you can make of it
>
> that is ... there is no particular PLAN to the investigation ... data
> gathering ... or, what you want to do with what you find after the fact
>
> i try to tell students this is  not a very good strategy ...

Dennis:
This is a serious misrepresentation of qualitative research in general, and, in
particular,  of Glaser and Strauss's insistence that social theorizing be
fundamentally be "grounded" in experience. The idea that qualitative research
begins with "no particular PLAN" is simply ludicrous.

Here's a couple of quotes from "Toward Reform of Program Evaluation" by Lee
Cronbach and others (1980):

 "The evaluator will be wise not to declare allegiance to either a
 quantitative-manipulative-summative methodology or a
 qualitative-naturalistic-descriptive methodology. . . . Those who
 advocate an evaluation plan [sic] devoid of one kind of information or
 the other carry the burden of justifying such exclusion." (p. 223)

 "Writings on experimental design emphasize extrapolation much less than
 is appropriate for evaluation. Data on a particular set of program
 operations are collected at specific sites, but the evaluation is
 intended to answer a much broader question."  . . . The design best for
 sharpening a limited statistical conclusion may not be the best basis
 for the broader inferences. " (p. 231)

Introductory statistics classes, with their artificially created null hypotheses
and impractical data gathering designs, often ignore these complexities.


--
  *
 `o^o' * Neil W. Henry ([EMAIL PROTECTED])   *
 -<:>- * Virginia Commonwealth University *
 _/ \_ * Richmond VA 23284-2014  *
  *  http://www.people.vcu.edu/~nhenry   *
  *


--9F3BD71D2EDA80683B2FF9EA
Content-Type: text/html; charset=us-ascii
Content-Transfer-Encoding: 7bit



 
dennis roberts wrote:
At 12:56 PM 3/25/01 -0500, Karl L. Wuensch wrote:

>Here is how I resolve that problem:  Define
the population from the sample,
>rather than vice versa -- that is, my results can be generalized to
any
>population for which my sample could be reasonably considered to be
a random
>sample.  Maybe we could call this "transcendental sampling" ;-)
-- it is
>somewhat like transcendental realism, defining reality from our percetion
of
>it, eh?
this sounds like  the method of grounded theory in the qualitative
bailiwick ...
look at data you have and see what you can make of it
that is ... there is no particular PLAN to the investigation ... data
gathering ... or, what you want to do with what you find after the
fact
i try to tell students this is  not a very good strategy ...
Dennis:
This is a serious misrepresentation of qualitative research in general,
and, in particular,  of Glaser and Strauss's insistence that social
theorizing be fundamentally be "grounded" in experience. The idea that
qualitative research begins with "no particular PLAN" is simply ludicrous.
Here's a couple of quotes from "Toward Reform of Program Evaluation"
by Lee Cronbach and others (1980):
"The evaluator will be wise not to declare allegiance to either
a quantitative-manipulative-summative methodology or a 
qualitative-naturalistic-descriptive
methodology. . . . Those who advocate an evaluation plan [sic] devoid of
one kind of information or the other carry the burden of justifying such
exclusion." (p. 223)
"Writings on experimental design emphasize extrapolation much less than
is appropriate for evaluation. Data on a particular set of program operations
are collected at specific sites, but the evaluation is intended to answer
a much broader question."  . . . The design best for sharpening a
limited statistical conclusion may not be the best basis for the broader
inferences. " (p. 231)
Introductory statistics classes, with their artificially created null hypoth

Re: Is there a test for causality?

2001-02-22 Thread Neil W. Henry



"G. Anthony Reina" wrote:

> Is there a test that variable X causes variable Y?

No, not in the abstract general way you pose the question

> I was under the impression that the best statistics could do was
> correlation not causation. In order to prove causation, one would have
> to know the specific mechanism whereby X could cause Y and possibly vary
> the input X to see if Y changed accordingly.

Under such a situation "statistics" comes into play in the analysis of the
data, which tend not to be able to speak for themselves.

>
> However, I've seen some papers on a method called 'directed coherence'
> which uses something called Granger causality. I think the basic gist is
> that the 'directed coherence' is the probability of predicting something
> about Y given you know something about X.
>

The phrase "Granger Causality" is used by econometricians, and derives from
a 1969 article in Econometrica by someone named --- you probably guessed
already --- Granger. The title will give you an idea of his approach:
"Investigating causal relations by econometric models and cross-spectral
methods."

Mainstream statisticians tend to ignore the extensive literature on
causation. It is almost never mentioned in textbooks even at the graduate
level. For an overview of  what some philosophers, historians and
statisticians have to say on the subject, look up

 Causality in crisis? : statistical methods and the search for
 causal knowledge in the social sciences edited by Vaughn R. McKim
 and Stephen P. Turner.
  Published:   Notre Dame, Ind. : University of Notre Dame Press,
 1997.
  Notes: Essays derived from presentations at the conference held
 in October 1993 at the University of Notre Dame.

The philosopher Clark Glymour and his collaborators claim to have developed
techniques for discovering causal structure in correlational data. Others
whose work you might look into include James Heckman ("selection modeling")
, Jamie Robins and Patrick Suppes ("probabilitic theory of
causality").

>
> Has anyone run across this?
>
> Thanks.
> -Tony Reina

--
  *
 `o^o' * Neil W. Henry ([EMAIL PROTECTED]) *
 -<:>- * Virginia Commonwealth University   *
 _/ \_ * Richmond VA 23284-2014  *
  *(804)828-1301 x124  *

  *FAX: 828-8785   http://www.people.vcu.edu/~nhenry *
  *




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=