I've been having issues similar to the ones others have mentioned with our
IRB. Also, our IRB has instituted a 2-week schedule for reviews (get it in
by x deadline and hear back by y date). Of course, they expect those
conducting research to meet deadlines, but they always run a week or so
behi
Here I quote from the material that has been used at many institutions for
training researchers to be IRB pure:
"If an evaluation systematically gathers data and contributes to
generalizable knowledge, or if the results are going to be published, the
activity is typically classified as research.
I'm surprised to hear that an IRB would make the dissemination of
results (in-class vs. public) the criterion for whether IRB review is
needed. Around here, the review issue hinges on who will be the
research participants. If students are only collecting data from each
other within the class,
I've been told that our IRB does not have jurisdiction over data
collection unless the "research" adds to the general body of scientific
knowledge. I guess that is why a classroom project that does not go outside
the classroom would be exempt. I have joked with my colleagues that were
this cr
Title: Position Announcement
Please forward this announcement to anyone who might be
interested.
Thank you.
Faith
---
Direct expanding NY State
Certificate Program in coordination with the business Human Resources
Program; teach a range of courses in industrial/organizational
psycho
A couple of weeks ago, I sent in a question about IRB problems and class
projects. Below are the responses I received. In defense of my local
IRB, they do allow surveys to be exempt provided in the results do not
go outside the classroom. This is true of experiments, too. In fact, I
may do any mini
Dear Rick,
a) standardized scores are no good solution since
it will make it impossible to tell the direction of pre-post
change, if there was any (recall that both set of scores will
have a mean of zero).
b) puting the two set of scores on the same metric (regardless of the
conversion method,
Hi Rick-
I have just gone through a similar exercise with a former student. she was
trying to compare a 5 and 8 point pre-post set of data. I suggested the
following:
Try thinking about it in a geometric, rather than an arithmatic way. Take a
piece of graph paper and draw out an X-Y graph with a
Rick Froman wrote:
> I have a colleague who wants to compare pretest performance on a 4-point scale with
> posttest performance on a 5 point scale. What is the best way to make such a
> comparison?
If your prediction is that the scores will go up, simply do a regular t-test between
the means o
Your colleague may have to create z-scores or T-scores (some
normalized measure) since the scales that were used are not
identical.
Bob Intrieri
At 12:00 PM 3/5/03 -0600, you wrote:
I have a colleague who wants to
compare pretest performance on a 4-point scale with posttest performance
on a 5 poin
Convert them both to 20-point scales.
*
Michael T. Scoles, Ph.D.
Director, Arkansas Charter School Resource Center
Associate Professor of Psychology & Counseling
University of Central Arkansas
Conway, AR 72035
voice: (501) 450-5418
fax:(501) 450
I have a colleague who wants to compare pretest performance on a 4-point scale with
posttest performance on a 5 point scale. What is the best way to make such a
comparison?
Rick
Dr. Rick Froman
Associate Professor of Psychology
John Brown University
Siloam Springs, AR 72761
(479) 524-7295
e-ma
"I often say that when you can measure what you are speaking about and
express it in numbers you know something about it; but when you cannot
measure it, when you cannot express it in numbers, your knowledge is of a
meager and unsatisfactory kind: it may be the beginning of knowledge, but
you have
Please forward this announcement to anyone who might be interested.
Thanks!
Rachelle Lipschultz
---
ANNE ARUNDEL COMMUNITY COLLEGE invites applications for non-tenure track,
term appointments, beginning Fall 2003, in a number of areas, including
Ps
James Guinee wrote:
> Paul Smith wrote:
> > But I still don't understand
> > what's going on here. Wouldn't "they do just as well based on outcome
data"
> > make those "theoretically-based approaches" into "empirically validated
> > approaches"?
>
> Well, let's say you used a theoretically-based a
> Subject: Re: Sternberg-Tavris
> From: "Paul Smith" <[EMAIL PROTECTED]>
> James Guinee wrote:
>
> > But some theoretically-based approaches are very difficult to test, and
> > Wampold is saying that they do JUST AS WELL based on outcome data
> > as the manualized stuff.
> After I sent my n
16 matches
Mail list logo