c for Splus under windows platform

2001-01-11 Thread Gang Liang

Hello all,

I'm looking for a c compiler which can work for Splus in windows. At least
gcc port for windows cygwin doesn't work, and I don't have watcom c. Anyone
has such experience before?

Thanks very much,
Gang




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Ronald Bloom

Herman Rubin <[EMAIL PROTECTED]> wrote:
> In article <[EMAIL PROTECTED]>,
> J. Williams <[EMAIL PROTECTED]> wrote:
>>Francis Galton explained it in 1885.  Possibly, the Mass. Dept. of
>>Education missed it!  Or, could it be that the same gang who brought
>>us the exit poll data during the November election were helping them
>>out?  :-)

>>I am wondering why they did not have a set of objective standards for
>>ALL  students to meet.

> There are only two ways this can be done.  One is by having
> the standards so low as to be useless, and the other is by
> not allowing the students who cannot do it to get to that
> grade, regardless of age.  The second is, at this time,
> Politically Incorrect.

  And what alternative do you propose?  Sending the underachievers
to work in the fields as soon as signs of promise fail
to manifest?  

 [...]

> The biggest factor in the performance of schools is in the
> native ability of students; but again it is Politically
> Incorrect to even hint that this differs between schools.

  It may be "politically incorrect" to say so.  But does that
support the proposition in any way shape or form?  So go 
on,  "hint"; get up on a beer-barrel and "hint" that the
"fit" are languishing from the ignominious condition of 
having to suffer the presence of the "unfit".  You'll
have plenty of company: Pride is greedier even than mere
Avarice.


-- R. Bloom



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Ronald Bloom

Robert J. MacG. Dawson <[EMAIL PROTECTED]> wrote:
>   
>   (2) _Why_ were even the best schools expected to improve, with targets
> that seem to have been overambitious?? I would hazard a guess that it
> might be due to an inappropriate overgeneralization of the philosophy -
> appropriate, in an eduational context, for individual students - that
> progress should constantly be being made. 

  Continual Growth is the state religion.  If you're not "growing" 
you're falling behind.  In business, there is no such thing anymore
as a "reasonable return" .  Nowadays, there is in it's stead the 
mantra of a reasonable rate of growth of the rate of return.



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Sample size for acceptance sampling

2001-01-11 Thread Eric Scharin

All -

I have a problem which I am hoping someone can help me solve.  (No, this is
not a problem from class, but a real-life problem from industry.)

A product has a specification of <= 1 defect/10 cm^2.  The product has an
overall surface area of X cm^2.  While this specification was not obtained
based on an analysis of the alpha & beta risks for acceptance sampling, the
specification itself implies an alpha & beta level.  Is there a simple way
to calculate how the alpha and beta risks would be affected by a change in
sample from 10 cm^2 to 5 cm^2?  Specifically, if someone accidentally
sampled 5 cm^2 and found 1 defect, what could you say about the status of
the product?

Thanks.

- Eric Scharin





=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Robert J. MacG. Dawson



Paul R Swank wrote:
> 
> Robert:
> 
> Why would you expect a strong correlation here? You're talking about tests done a 
>year apart with some new kids in each school and some kids who have moved on.

Simply because there seems to be general consensus that there are such
things as "good schools" and "poor schools".  This may be partially
because some schools have catchment areas in which parents are better
educated/more suppportive/able to afford breakfast for their kids and
others aren't, partially because schools in some areas have better
funding,  partially because some schools have better teachers, or for a
host of other reasons. 


> Is regression toward the mean causing all of the noted results. Probably not. But it 
>is quite conceivable that it could be partially responsible for the results.

It would certainly contribute - but I think it would be a minor
contribution, in the presence of goals as stated.

-Robert Dawson


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Rich Ulrich

On Wed, 10 Jan 2001 21:32:43 GMT, Gene Gallagher
<[EMAIL PROTECTED]> wrote:

> The Massachusetts Dept. of Education committed what appears to be a
> howling statistical blunder yesterday.  It would be funny if not for the
> millions of dollars, thousands of hours of work, and thousands of
> students' lives that could be affected.
> 
< snip, much detail > 
> I find this really disturbing.  I am not a big fan of standardized
> testing, but if the state is going to spend millions of dollars
> implementing a state-wide testing program, then the evaluation process
> must be statistically valid.  This evaluation plan, falling prey to the
> regression fallacy, could not have been reviewed by a competent
> statistician.
> 
> I hate to be completely negative about this.  I'm assuming that
> psychologists and others involved in repeated testing must have
> solutions to this test-retest problem.

The proper starting point for a comparison for a school
should be the estimate of the "true score" for the school:
the regressed-predicted value under circumstances of
no-change.  "No-change" at the bottom would be satisfied
by becoming only a little better; no-change at the top 
would be met by becoming only a little worse.  If you are 
dumping money into the whole system, then you might 
hope to (expect to?) bias the changes into a positive direction.

I thought it was curious that the 
  "schools in the highest two categories were expected to increase
their average MCAS scores by 1 to 2 points, while schools in the
lowest two categories were expected to improve their scores by 4-7
points."  

That sounds rational in form.  It appears to me that their model 
might have the correct form, but the numbers surprised them.
That is:  It looks as if someone was taking into account regression of
a couple of points, then hoping for a gain of 4 or 5 points.  That
(probably) under-estimated the regression-to-the-mean, and
over-estimated how much a school could achieve by freshly 
swearing to good intentions.

What is needed -- in addition to self-serving excuses -- is an
external source of validation.  And it should validate in cases that
are not predicted by regression to the mean.

-- 
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Number of classes.

2001-01-11 Thread Jon Cryer

I asked Minitab support how they did it. Here is their answer:

Date: Fri, 26 Sep 1997 15:07:50 -0400
To: [EMAIL PROTECTED]
From: Tech Support 
Subject: number of bars in MINITAB histogram

Jonathan,

I finally found an answer for you.  Here's the algorithm.

There are upper and lower bounds on the number of bars.

Lower bound = Round( (16.0*N)**(1.0/3.0) + 0.5 )
Upper bound = Lower bound + Round(0.5*N)

After you find the bounds, MINITAB will always try to get as close to the
lower bound as it can.

Then we have a "nice numbers" algorithm that finds interval midpoints,
given the constraints on the number of intervals.

But there is special code for date/time data and for highly granular data
(e.g., all 1's and 2's).

Find the largest integer p such that each data value can be written (within
fuzz) as an integer times 10**p.

Let BinWidth = 10**p.

Let BinCount =  1 + Round( ( range of data ) / BinWidth )

If BinCount  is <= 10, then let the bin midpoints run from the data min to
the data max in increments of BinWidth.

Otherwise, use the "nice numbers" algorithm.

Hope this helps.

Andy Haines
Minitab, Inc.

At 11:01 PM 1/4/01 -0500, you wrote:
>To determine the number of classes for a histogram, Excel uses square root
>of the number of observations. Is it also true for the number of
>observations greater than 200, say, for 2000?. Does the MINITAB use the same
>for determining the number of classes for a histogram?
>Any help would be appreciated.
>
>
>
>
>
>=
>Instructions for joining and leaving this list and remarks about
>the problem of INAPPROPRIATE MESSAGES are available at
>  http://jse.stat.ncsu.edu/
>=
>
>
 ___
--- |   \
Jon Cryer, Professor [EMAIL PROTECTED]   ( )
Dept. of Statistics  www.stat.uiowa.edu/~jcryer \\_University
 and Actuarial Science   office 319-335-0819 \ *   \of Iowa
The University of Iowa   dept.  319-335-0706  \/Hawkeyes
Iowa City, IA 52242  FAX319-335-3017   |__ )
---   V

"It ain't so much the things we don't know that get us into trouble. 
It's the things we do know that just ain't so." --Artemus Ward 


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Paul R Swank
Robert:

Why would you expect a strong correlation here? You're talking about tests done a year apart with some new kids in each school and some kids who have moved on.
Is regression toward the mean causing all of the noted results. Probably not. But it is quite conceivable that it could be partially responsible for the results. 

At 03:21 PM 1/11/01 -0400, you wrote:
>
>
>Paul R Swank wrote:
>> 
>> Regression toward the mean occurs when the pretest is used to form the groups, which it appears is the case here.
>
>	Of course it "occurs": - but remember that the magnitude depends on
>r^2. In the case where there is strong correlation between the pretest
>and the posttest, we do not expect regression to the mean to be
>particularly significant. 
>
>	Now, it is generally acknowledged that there are some schools which
>_consistently_ perform better than others. (If that were not the case,
>nobody would be much surprised by any one school failing to meet its
>goal!)  Year-over-year variation for one school is presumably much less
>than between-school variation. 
>
> 	Therefore, I would not expect regression to the mean to be sufficient
>to explain the observed outcome (in which "practically no" top schools
>met expectations); and I conclude that the goals may well have been
>otherwise unreasonable. Indeed, requiring every school to improve its
>average by at least two points every year is not maintainable in the
>long run, and only justifiable in the short term if there is reason to
>believe that *all* schools are underperforming. 
>
>	-Robert Dawson
>
>
>=
>Instructions for joining and leaving this list and remarks about
>the problem of INAPPROPRIATE MESSAGES are available at
>  http://jse.stat.ncsu.edu/
>=
>

Paul R. Swank, PhD.
Professor & Advanced Quantitative Methodologist
UT-Houston School of Nursing
Center for Nursing Research
Phone (713)500-2031
Fax (713) 500-2033

= Instructions for joining and leaving this list and remarks about the problem of INAPPROPRIATE MESSAGES are available at http://jse.stat.ncsu.edu/ = 

Re: MA MCAS statistical fallacy

2001-01-11 Thread Herman Rubin

In article <[EMAIL PROTECTED]>,
J. Williams <[EMAIL PROTECTED]> wrote:
>Francis Galton explained it in 1885.  Possibly, the Mass. Dept. of
>Education missed it!  Or, could it be that the same gang who brought
>us the exit poll data during the November election were helping them
>out?  :-)

>I am wondering why they did not have a set of objective standards for
>ALL  students to meet.

There are only two ways this can be done.  One is by having
the standards so low as to be useless, and the other is by
not allowing the students who cannot do it to get to that
grade, regardless of age.  The second is, at this time,
Politically Incorrect.



When a given school is already "good" it naturally can't
>"improve" more than schools on the bottom of the achievement ladder.

It can, by changing curriculum and speeding things up.  This
is also not Politically Correct.

>It seems they really should have prepared a better public announcement
>of results.  Rather than "knocking" the high achieving schools, they
>should praise them justifiably.  Then, noting the improvement in the
>large urban schools would seem positive as well.

The biggest factor in the performance of schools is in the
native ability of students; but again it is Politically
Incorrect to even hint that this differs between schools.
-- 
This address is for information only.  I do not claim that these views
are those of the Statistics Department or of Purdue University.
Herman Rubin, Dept. of Statistics, Purdue Univ., West Lafayette IN47907-1399
[EMAIL PROTECTED] Phone: (765)494-6054   FAX: (765)494-0558


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Robert J. MacG. Dawson



Paul R Swank wrote:
> 
> Regression toward the mean occurs when the pretest is used to form the groups, which 
>it appears is the case here.

Of course it "occurs": - but remember that the magnitude depends on
r^2. In the case where there is strong correlation between the pretest
and the posttest, we do not expect regression to the mean to be
particularly significant. 

Now, it is generally acknowledged that there are some schools which
_consistently_ perform better than others. (If that were not the case,
nobody would be much surprised by any one school failing to meet its
goal!)  Year-over-year variation for one school is presumably much less
than between-school variation. 

Therefore, I would not expect regression to the mean to be sufficient
to explain the observed outcome (in which "practically no" top schools
met expectations); and I conclude that the goals may well have been
otherwise unreasonable. Indeed, requiring every school to improve its
average by at least two points every year is not maintainable in the
long run, and only justifiable in the short term if there is reason to
believe that *all* schools are underperforming. 

-Robert Dawson


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



MCAS

2001-01-11 Thread Alan Zaslavsky

Dear Gene,

I share your feelings about MCAS (fortunately my daughter finished high
school before it came into effect, but that's no consolation for the
thousands of other kids who are supposed to be deprived of their high
school diplomas), and had some similar reactions to the article.

On the statistical issue, there are two ways that this could be done that
have different implications.  If you form categories based on the actual
score in the previous year, then you have a classic regression to the
mean situation.  The magnitude of the regression to the mean effect could
be estimated knowing variances and sample sizes at the various schools;
it is even possible that allowing for a few points smaller required
advance in the higher-scoring school might adjust for that effect
(although I would be surprised!).  

Another approach would be to define groups by some other variable related
to but distinct from outcomes, e.g. inner city versus suburbs, percent
minority, or percent in poverty.  Although the groups will differ in their
baseline values, there is no regression to the mean effect there.

However, either of these analyses begs the question of the potential for
improvement in different schools.  An excellent school may already be doing
everything it should be doing and have no way to improve, while a low-scoring
school may have a lot of possible avenues to improvement.  There may also
be issues about the appropriateness of the educational strategies that might
be adopted.  Somebody might argue that at a school already doing well at
MCAS, improvements could only be obtained by teaching to the MCAS at the
cost of less investment in AP exams.  (I personally have little confidence
in the educational incentives of the MCAS at any level; I am just laying
out possible arguments.)  Presumably some arguments could also be advanced
to the opposite effect, as well.

Best regards
Alan Zaslavsky

P.S. I see now that Robert Dawson had already made some similar comments,
which I hadn't gotten to yet when I wrote this.


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread dennis roberts

i went to some of the sites given in the urls ... and, quite frankly, it is 
kind of difficult to really get a feel for what has transpired ... and how 
targets were set ... and how goals were assessed

regardless of whether we like this kind of an approach for accountability 
... or not ... we all have to admit that there are a host of problems with 
it ... many of these are simply political and policy oriented (in fact, 
these might be the largest of the problems ... when the legislature starts 
enacting regulations without a real good understanding of the 
methodological problems)  ... some are measurement related ... and yes, 
some are statistical in nature

we do have components of this total process

1. there are tests that are developed/used/administered/scored ... in 4th 
and 8th and 10th grades ... these are NOT the same tests of course ... so, 
one is never sure what it means to compare "results" from say the 8th grade 
to the 4th grade ... etc.

2. then we have the problem that one year ... we have the data on the 4th 
graders THAT year ... but, the next year we have data on the 4th graders 
for THAT year ... these are not the same students ... so any direct 
comparison of the scores ... 4th to 4th ... or 8th to 8th or 10th to 10th 
... are not totally comparable ... so, ANY difference in the scores ... up 
or down ... cannot be necessarily attributed to improvement or lack of 
improvement ... the changes could be related and in fact, totally accounted 
for because there are small changes in the abilities of the 4th graders one 
year compared to another year ... (or many other reasons)

3. as i said before, we also have the problem of using aggregated 
performance ... either averages of schools and/or averages for districts 
... when we line them up and then assign these quality names of very high, 
high, etc.
there is a necessary disconnect between the real goals of education ... 
that is, helping individual kids learn .. and the way schools or districts 
are being evaluated ... when averages are being used ...

4. i would like to know how on earth ... these 'standards' for 
"dictated"  improvement targets were derived ... did these have anything to 
do with real data ... or an analysis of data ... or, were just roundtabled 
and agreed to? we have to know this to be able to see if there is any 
connection between policy targets and statistical problems

5. we have to try to separate relative standing data from actual test score 
gain information ... and we don't know how or if the ones setting the 
standards and making decisions ... know anything about this problem

so, to summarize ... there are many many issues and problems with 
implementing any system whereby you are trying to evaluate the performance 
of schools and districts ... and, perhaps the least important of these is a 
statistical one ... set in the context of political policy matters ... that 
a legislature works with ... and "legislates" targets and practices without 
really understanding the full gamut of difficulties when doing so

unfortunately, in approaches like this, one makes an assumption that if a 
school ... or district ... gets better (by whatever measure) ... that this 
means that individual students get better too ... and we all know of course 
that this is NOT NECESSARILY TRUE ... and in fact we know more than that 
... we know that it is NOT true ... in many cases

sure, it is important to make some judgements about how schools and 
districts are doing ... especially if each gets large sums of money from 
taxpayers ... but, the real issue is how we work with students ... and how 
each and every one of them do ... how each kid improves or not ... and, all 
these approaches to evaluating schools and districts ... fail to keep that 
in mind ... thus, in the final analysis, all of these systems are 
fundamentally flawed ... (though they still may be useful)




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread dennis roberts

At 11:31 PM 1/10/01 -0500, Bob Hayden wrote:

regression to the mean applies to relative position ... NOT raw scores

let's say we give a test called a final exam at the beginning of a course 
... and assume for a moment that there is some spread ... though the mean 
necessarily would be rather low ... then, we give an alternate form of this 
final exam at the end of the course ... where again, there is reasonable 
spread but, obviously, the mean has gone up alot ...

now, EVERYONE'S SCORES GO UP ... so everyone improves ... and it is not 
that the low scores (BECAUSE of regression) will improve more and the 
better scoring students on the pretest will improve less)  that is NOT what 
regression to the mean is all about ...

so, it depends on how these tests are scored and reported ... if the scores 
are reported on something like a percentile  score basis ... then there is 
necessarily a problem ... but, if the scores are reported on some scale 
that reflect that 10th grade scores are higher than 8th grade scores ... 
and 8th grade scores are necessarily higher than 4th grade scores ... that 
is, the scores reflect an ever increasing general level of knowledge ... 
then regression to the mean is not the bugaboo that the "letter" makes it 
out to be

now, the post said:

The effectiveness of school districts is being assessed using average
student MCAS scores. Based on the 1998 MCAS scores, districts were
placed in one of 6 categories: very high, high, moderate, low, very low,
or critically low. Schools were given improvement targets based on the
1998 scores, with schools in the highest two categories were expected to
increase their average MCAS scores by 1 to 2 points, while schools in
the lowest two categories were expected to improve their scores by 4-7
points

=
there are a number of ? that this paragraph brings to mind:

1. how are categories of very high, etc. ... translated into 1 to 2 points 
... or 4 to 7 points? i don't see any particular connection of one to the other

2. we have a problem here of course that the scores in a district are 
averages ... not scores for individual kids in 4th, 8th, and 10th grades

3. what does passing mean in this context?

4. let's say there are 50 districts ... and, for last  year ... using 4th 
grade as an example ... we line up from highest mean for a district down to 
lowest mean for a district  then, in the adjacent column, we put what 
those same districts got as means on the tests for the 4th grade this year 


we would expect this correlation to be very high ... for two reasons ... 
first, means are being used and second, from year to year ... school 
district's population does not change much ... so if one district has on 
average, a lower scoring group of 4th grade students  that is what is 
going to be the case next year

thus, given this ... we would NOT expect there to be much regression to the 
mean ... since the r between these two variables i bet is very high

5. but, whatever the case is in #4 ... what does this have to do with 
CHANGE IN MEAN SCORES? or changes in the top group of at least 1/2 points 
and in the low groups changes of 4/7 points? the lack of r between the two 
years of 4th grade means on these test just means that their relative 
positions change with the higher ones not looking as relatively high ... 
and the low ones not looking quite so relatively low BUT, your position 
could change up or down relatively speaking regardless of whether your mean 
test performance went up or down ... or stayed the same


bottom line: we need alot more information about exactly what was done ... 
and how improvement goals were defined in the first place ... before we can 
make any reasonable inference that regression to the mean would have 
anything to do with better districts being bad mouthed and poorer 
performing districts being praised



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Paul R Swank
Regression toward the mean occurs when the pretest is used to form the groups, which it appears is the case here.

At 08:31 AM 1/11/01 -0400, you wrote:
>
>
>Gene Gallagher wrote:
>> 
>> Those familiar with "regression to the mean" know what's coming next.
>> The poor schools, many in urban centers like Boston, met their
>		^
>> improvement "targets," while most of the state's top school districts
>> failed to meet their improvement targets.
>
>	Wait one... Regression to the mean occurs because of the _random_
>component in the first measurement. Being in an urban center is not part
>of the random component - those schools' grades didn't improve because 
>some of them woke up one day and found that their school had moved to a 
>wealthier district.
>	If the effect of nonrandom components such as this is large enough 
>(as I can well believe) to justify the generalization highlighted above,
>and if there was a strong pattern of poor-performing schools meeting
>their
>targets and better-performing schools not doing so, we are looking at
>something else - what, I'll suggest later
>
>
>> The Globe article describes how superintendents of high performing
>> school districts were outraged with their failing grades, while the
>> superintendent of the Boston school district was all too pleased with
>> the evaluation that many of his low-performing schools had improved:
>> 
>> [Brookline High School, for example, with 18 National Merit Scholarship
>> finalists and the highest SAT scores in  years, missed its test-score
>> target - a characterization  blasted by Brookline Schools Superintendent
>> James F. Walsh, who dismissed the report.
>
>	There *is* a problem here,but it's not (entirely) regression
>to the mean. If I recall correctly, Brookline High School is
>internationally
>known as an excellent school, on the basis of decades of excellent
>teaching.
>If it couldn't meet its target, it's not because its presence among the
>top
>schools was a fluke in the first measurement - it's probably because the 
>targets for the top schools were unrealistic.
>
>	Was there any justification for the assumption voiced by the Boston
>superintendant that the top-performing schools were in fact not
>performing at
>their capacity and would be "smug" if they assumed that their present
>per-
>formance was acceptable?  The targets described seem to imply that no
>school
>in the entire state - not one - was performing satisfactorily, even the
>top 
>ones. Perhaps this was felt to be true, or perhaps it was politically
>more
>acceptable to say "you all need to pull your socks up" than to say "the 
>following schools need to pull their socks up; the rest of you, steady
>as she goes." 
>
>	As a reductio ad absurdum, if this policy were followed repeatedly,
>it would be mathematically impossible for any school to meet its target
> every year. That - and not regression to the mean - is the problem
>here, I
>think.
>
>		-Robert Dawson
>
>
>=
>Instructions for joining and leaving this list and remarks about
>the problem of INAPPROPRIATE MESSAGES are available at
>  http://jse.stat.ncsu.edu/
>=
>

Paul R. Swank, PhD.
Professor & Advanced Quantitative Methodologist
UT-Houston School of Nursing
Center for Nursing Research
Phone (713)500-2031
Fax (713) 500-2033

= Instructions for joining and leaving this list and remarks about the problem of INAPPROPRIATE MESSAGES are available at http://jse.stat.ncsu.edu/ = 

Re: MA MCAS statistical fallacy

2001-01-11 Thread J. Williams

Francis Galton explained it in 1885.  Possibly, the Mass. Dept. of
Education missed it!  Or, could it be that the same gang who brought
us the exit poll data during the November election were helping them
out?  :-)

I am wondering why they did not have a set of objective standards for
ALL  students to meet.  Of course, it is nice to reward academically
weaker  districts for "improving,"  but the real issue may not be
"improvement,"  rather it might be attainment at a specific level for
all schools as a minimum target.  A sliding scale depicting
"improvement" means little if the schools in question are producting
students who fall behind in math, reading comprehension, etc.
Rewarding urban schools for improving probably is a good idea, but
that should not mean entering  a zero sum game with the "good"
schools.  When a given school is already "good" it naturally can't
"improve" more than schools on the bottom of the achievement ladder.
It seems they really should have prepared a better public announcement
of results.  Rather than "knocking" the high achieving schools, they
should praise them justifiably.  Then, noting the improvement in the
large urban schools would seem positive as well.


On Wed, 10 Jan 2001 21:32:43 GMT, Gene Gallagher
<[EMAIL PROTECTED]> wrote:

>The Massachusetts Dept. of Education committed what appears to be a
>howling statistical blunder yesterday.  It would be funny if not for the
>millions of dollars, thousands of hours of work, and thousands of
>students' lives that could be affected.
>
>Massachusetts has implemented a state-wide mandatory student testing
>program, called the MCAS.  Students in the 4th, 8th and 10th grades are
>being tested and next year 12th grade students must pass the MCAS to
>graduate.
>
>The effectiveness of school districts is being assessed using average
>student MCAS scores.  Based on the 1998 MCAS scores, districts were
>placed in one of 6 categories: very high, high, moderate, low, very low,
>or critically low.  Schools were given improvement targets based on the
>1998 scores, with schools in the highest two categories were expected to
>increase their average MCAS scores by 1 to 2 points, while schools in
>the lowest two categories were expected to improve their scores by 4-7
>points (http://www.doe.mass.edu/ata/ratings00/rateguide00.pdf).
>
>Based on the average of 1999 and 2000 scores, each district was
>evaluated yesterday on whether they had met their goals.  The report was
>posted on the MA Dept. of education web site:
>http://www.doe.mass.edu/news/news.asp?id=174
>
>Those familiar with "regression to the mean" know what's coming next.
>The poor schools, many in urban centers like Boston, met their
>improvement "targets," while most of the state's top school districts
>failed to meet their improvement targets.
>
>The Boston Globe carried the report card and the response as a
>front-page story today:
>http://www.boston.com/dailyglobe2/010/metro/Some_top_scoring_schools_fau
>lted+.shtml
>
>The Globe article describes how superintendents of high performing
>school districts were outraged with their failing grades, while the
>superintendent of the Boston school district was all too pleased with
>the evaluation that many of his low-performing schools had improved:
>
>[Brookline High School, for example, with 18 National Merit Scholarship
>finalists and the highest SAT scores in  years, missed its test-score
>target - a characterization  blasted by Brookline Schools Superintendent
>James F. Walsh, who dismissed the report.
>
>"This is not only not helpful, it's bizarre," Walsh said.  ''To call
>Brookline, Newton, Medfield, Weston, Wayland, Wellesley as failing to
>improve means so little, it's not helpful. It becomes absurd when you're
>using this formula the way they're using it.''
>
>Boston School Superintendent Thomas W. Payzant, whose district had 52 of
>113 schools meet or exceed expectations, was more blunt: "For the
>high-flying schools, I say they have a responsibility to not be smug
>about the level they have reached and continue to aspire to do better."]
>
>
>
>Freedman, Pisani & Purvis (1998, Statistics 3rd edition) describe the
>fallacy involved:
>"In virtually all test-retest situations, the bottom group on the first
>test will on average show some improvement on the second test and the
>top group will on average fall back.  This is the regression effect.
>Thinking that the regression effect must be due to something important,
>..., is the regression fallacy."
>
>I find this really disturbing.  I am not a big fan of standardized
>testing, but if the state is going to spend millions of dollars
>implementing a state-wide testing program, then the evaluation process
>must be statistically valid.  This evaluation plan, falling prey to the
>regression fallacy, could not have been reviewed by a competent
>statistician.
>
>I hate to be completely negative about this.  I'm assuming that
>psychologists and others involved in repeated testing must hav

Re: MA MCAS statistical fallacy

2001-01-11 Thread Robert J. MacG. Dawson

A couple additional thoghts I didn't get around to before leaving for
my 8:30 lecture:

(1) The clearest way of looking at the stats side of things is probably
that one would expect a high enough r^2 between schools' performances in
one year and in the next that regression to the mean would be a rather
minor phenomenon.

(2) _Why_ were even the best schools expected to improve, with targets
that seem to have been overambitious?? I would hazard a guess that it
might be due to an inappropriate overgeneralization of the philosophy -
appropriate, in an eduational context, for individual students - that
progress should constantly be being made. 

Getting further off-topic, we see the same thing in economics, where
the standard model for the Western economies is one of constant growth,
and our institutions seem unable to adapt to slight shrinkage - or even
slower-than-usual growth - without pain all round. 

Our culture seems to concentrate on the idea that the moment an
institution stops growing it starts to die, and does not seem to put
much effort into maintaining "mature" institutions for which the
constant-growth paradigm is no longer appropriate.  Maybe this cultural 
neoteny is still appropriate and advantageous at this stage in history -
I don't know. 


-Robert


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MA MCAS statistical fallacy

2001-01-11 Thread Robert J. MacG. Dawson



Gene Gallagher wrote:
> 
> Those familiar with "regression to the mean" know what's coming next.
> The poor schools, many in urban centers like Boston, met their
^
> improvement "targets," while most of the state's top school districts
> failed to meet their improvement targets.

Wait one... Regression to the mean occurs because of the _random_
component in the first measurement. Being in an urban center is not part
of the random component - those schools' grades didn't improve because 
some of them woke up one day and found that their school had moved to a 
wealthier district.
If the effect of nonrandom components such as this is large enough 
(as I can well believe) to justify the generalization highlighted above,
and if there was a strong pattern of poor-performing schools meeting
their
targets and better-performing schools not doing so, we are looking at
something else - what, I'll suggest later


> The Globe article describes how superintendents of high performing
> school districts were outraged with their failing grades, while the
> superintendent of the Boston school district was all too pleased with
> the evaluation that many of his low-performing schools had improved:
> 
> [Brookline High School, for example, with 18 National Merit Scholarship
> finalists and the highest SAT scores in  years, missed its test-score
> target - a characterization  blasted by Brookline Schools Superintendent
> James F. Walsh, who dismissed the report.

There *is* a problem here,but it's not (entirely) regression
to the mean. If I recall correctly, Brookline High School is
internationally
known as an excellent school, on the basis of decades of excellent
teaching.
If it couldn't meet its target, it's not because its presence among the
top
schools was a fluke in the first measurement - it's probably because the 
targets for the top schools were unrealistic.

Was there any justification for the assumption voiced by the Boston
superintendant that the top-performing schools were in fact not
performing at
their capacity and would be "smug" if they assumed that their present
per-
formance was acceptable?  The targets described seem to imply that no
school
in the entire state - not one - was performing satisfactorily, even the
top 
ones. Perhaps this was felt to be true, or perhaps it was politically
more
acceptable to say "you all need to pull your socks up" than to say "the 
following schools need to pull their socks up; the rest of you, steady
as she goes." 

As a reductio ad absurdum, if this policy were followed repeatedly,
it would be mathematically impossible for any school to meet its target
 every year. That - and not regression to the mean - is the problem
here, I
think.

-Robert Dawson


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



? re influence of cor. coeff.

2001-01-11 Thread Dr. S. Shapiro

Dear Colleagues;

I derived the following equation using MINITAB:

pKi = - 10.9 + 3.77 Tu + 1.89 Tm + 5.55 Av 
  - 6.35 Ap + 37.0 Km - 71.6 Ks

Predictor   Coef   StDev  TP   VIF
Constant -10.894   2.798  -3.890.005
Tu3.7654  0.6885   5.470.000  37.6
Tm1.8904  0.4559   4.150.003  21.0
Av 5.550   1.100   5.050.0002187.1
Ap-6.353   1.146  -5.540.0002641.9
Km37.034   8.609   4.300.003  34.7
Ks-71.61   14.21  -5.040.000  85.5

n = 15  SD = 0.5804  R-Sq = 89.6%  F = 11.51


Since the variance inflation factors are so large I generated a
Pearson correlation maxtix:

Tu  Tm   Av   Ap   Km
Tm0.261
Av0.7790.405
Ap0.7720.4300.999
Km   -0.1730.525   -0.436   -0.415
Ks   -0.0360.368   -0.474   -0.4630.931
 

As you see Ap and Av are colinear.  So I thought to see what
would happen if either of these two variables were omitted from
the equation.  When the data were regressed without Ap I
obtained

pKi = - 5.23 + 1.67 Tu + 0.284 Tm - 0.473 Av + 6.8 Km - 17.6 Ks

Predictor   Coef   StDev  TP   VIF
Constant  -5.228   5.404  -0.970.359
Tu 1.672   1.194   1.400.195  26.3
Tm0.2837  0.7302   0.390.707  12.5
Av   -0.4735  0.3596  -1.320.220  54.3
Km  6.79   13.82   0.490.635  20.8
Ks-17.63   21.49  -0.820.433  45.4

n = 15  S = 1.204   R-Sq = 49.7%  F = 1.78


whereas is Av is omitted I obtain

pKi = - 7.37 + 2.32 Tu + 0.694 Tm - 0.645 Ap + 11.3 Km - 29.9 Ks

Predictor   Coef   StDev  TP   VIF
Constant  -7.366   5.223  -1.410.192
Tu 2.324   1.208   1.920.086  31.2
Tm0.6935  0.7505   0.920.380  15.3
Ap   -0.6447  0.3480  -1.850.097  65.6
Km 11.25   13.36   0.840.421  22.5
Ks-29.94   22.30  -1.340.212  56.6

n = 15  S = 1.119   R-Sq = 56.6%  F = 2.34


My question is: why is there such a decline in the statistical
quality of the regression equation when one of the two colinear
variables (Av or Ap) is omitted??  Since Ap and Av are colinear,
I expected that removing one of them from the 6-variable
equation would have been compensated for by a change in the
coefficient preceding the other, accompanied by only a
negligible change in statistical quality of the 5-variable
equations compared to the 6-variable equation.  Obvious this was
_not_ the case.

Responders should contact me _directly_ at

[EMAIL PROTECTED]

because I rarely log on to this usegroup.

Thanks in advance to all responders,

S. Shapiro
[EMAIL PROTECTED]



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=