Well-put!  It would be great if people (particularly reviewers) always kept
this in mind.

--Ruchira

On Sun, Mar 20, 2011 at 7:57 AM, James J. Roper <jjro...@gmail.com> wrote:

> I've been meaning to comment here too.
>
> When I teach statistics, my goal is to give the graduate students a
> "toolbox" if you will, of useful ways to test ideas.  More complex
> statistics comes later.  In teaching, I use the idea of testing hypotheses,
> with a very important caveat.  Both, null and alternative hypotheses have
> to
> be biologically sensible and biologically possible.  I know I find many
> published papers that gloss over the null, but it turns out, on deeper
> inspection, that it was not a possibility and so refuting it was
> unavoidable.
>
> Apply that idea, that the null also must be reasonable, logical and
> possible, and you may find that many null hypotheses are none of those.
>
> Jim
>
> On Thu, Mar 10, 2011 at 13:13, Kevin Mueller <kem...@psu.edu> wrote:
>
> > If we iteratively modify our hypotheses through the process of data
> > collection, data analysis, or manuscript preparation, how different is
> this
> > process from "observational" or "exploratory" research?  It is, of
> course,
> > different to some debatable extent. Regardless, I think Paul's comments
> shed
> > light on the reality that there is a large gray area between the extremes
> of
> > purely observational studies and purely hypothesis driven studies (which
> his
> > 2005 paper apparently documents).  Given this, I find the explicit or
> > underlying claims of superiority made by proponents of hypothesis driven
> > research to ring false (despite some of the strong benefits of hypothesis
> > testing that Paul and others have made clear).  I find such claims ironic
> > since the result of many observational or exploratory studies is, gasp, a
> > hypothesis.
> >
> > Finally, regardless of the language we use to reference hypotheses in our
> > introductions, I ask:  Is it always beneficial to cloak studies that are
> > somewhat exploratory behind a veil of singlular hypothesis testing?  Or
> > might we also sometimes gain and share insights by making the process of
> > data exploration and hypothesis testing/modification more apparent in our
> > manuscripts?
> >
> > To be clear, my comments are more in response to a general
> > narrow-mindedness that I've observed among some natural scientists, not
> to
> > any particular post or 'poster' in this recent thread (i.e. I found
> Paul's
> > post insightful and not especially narrow-minded).
> >
> > Kevin Mueller
> >
> > On Mar 9, 2011, at 11:00 PM, Paul Grogan wrote:
> >
> > "Furthermore, often during the data interpretation or write-up
> > stage, additional reflection on the processes of experimentation and
> > evaluation of the data may indicate to the scientist (or to a manuscript
> > reviewer) that the test did not reflect the hypothesis as well as
> > originally
> > thought. In such cases, further refinement or editing of the hypothesis
> > statement should be made so that the final research output – the
> > peer-reviewed publication disseminating the new knowledge – is as
> accurate
> > and accessible to others as possible.  As a result, I usually finish my
> > manuscript Introduction sections with: “We used our data to test the
> > following hypotheses....” (rather than “We tested the following
> > hypotheses... which gives the impression of great foresight on the part
> of
> > the author)."
>

Reply via email to