Re: adjusted r-square

2001-08-22 Thread Graeme Byrne

In short, you don't. If the number of terms in the model equals the number
of observations you have much bigger problems than not being able to compute
adjusted R^2. It should always be the case that the number of observations
exceed the number of terms in the model otherwise you cannot calculate any
of the standard regression diagnostics (F-stats, t-stats etc). My advice is
get more data or remove terms from the model. If neither of these is an
option you are stuck.


Atul [EMAIL PROTECTED] wrote in message
[EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
 I have a doubt regarding adjusted r-square

 How do we calculate the adjusted r-square when the error degrees of
 freedom are zero ?
 (or in other words, number of samples is equal  to the number of
 regression terms including the constant)

 Such a situation leads to a zero in the denominator in the expression
 for calculating adjusted r-square.

 Your help is highly appreciated.

 Thanks
 Atul




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: adjusted r-square

2001-08-22 Thread Eric Bohlman

In sci.stat.consult Graeme Byrne [EMAIL PROTECTED] wrote:
 In short, you don't. If the number of terms in the model equals the number
 of observations you have much bigger problems than not being able to compute
 adjusted R^2. It should always be the case that the number of observations
 exceed the number of terms in the model otherwise you cannot calculate any
 of the standard regression diagnostics (F-stats, t-stats etc). My advice is
 get more data or remove terms from the model. If neither of these is an
 option you are stuck.

It's worse than not being able to calculate regression diagnostics.  You 
can't make *any* inferences beyond your observed data.  Consider the 
degenerate case of trying to fit a bivariate regression line when you have 
only two observations.  You'll *always* get a perfect fit because two 
points mathematically define a line.  But that perfect fit will tell you 
absolutely nothing about the underlying relationship between the two 
variables.  It's consistent with *any* possible relationship, including 
complete independence.  You can't tell how far off your model is from the 
observations because there simply isn't any room for it to be off (room 
for the model to be off otherwise being known as degrees of freedom).

A model with as many parameters as observations is equivalent to the 
observations themselves, and therefore testing such a model against the 
observations is the same thing as asking if the observations are equal to 
themselves, which is circular reasoning.



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: adjusted r-square

2001-08-22 Thread Joe Ward

If the least-squares regression algorithm does not
REQUIRE THE NUMBER OF OBSERVATIONS TO EXCEED
THE NUMBER OF PREDICTORS, THEN THE REGRESSION
ALGORITHM COULD BE USED TO SOLVE A SYSTEM OF
SIMULTANEOUS EQUATIONS THAT WOULD HAVE
NO ERRORS.

Another interesting characteristic of Excel Regression is that it
requires
the number of observations to exceed the number of predictors.

Fortunately, Colin Bell is working with the Excel folks at Microsoft to
improve the numerous interesting characteristics of  Statistics in Excel.

-- Joe



*** Joe H. Ward,  Jr.
*** 167 East Arrowhead Dr.
*** San Antonio, TX 78228-2402
*** Phone: 210-433-6575
*** Fax:   210-433-2828
*** Email: [EMAIL PROTECTED]
*** http://www.ijoa.org/resumes/ward.html
*** ---
*** Health Careers High School
*** 4646 Hamilton-Wolfe
*** San Antonio, TX 78229
*




- Original Message -
From: Graeme Byrne [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Sent: Wednesday, August 22, 2001 4:42 AM
Subject: Re: adjusted r-square


 In short, you don't. If the number of terms in the model equals the number
 of observations you have much bigger problems than not being able to
compute
 adjusted R^2. It should always be the case that the number of observations
 exceed the number of terms in the model otherwise you cannot calculate any
 of the standard regression diagnostics (F-stats, t-stats etc). My advice
is
 get more data or remove terms from the model. If neither of these is an
 option you are stuck.


 Atul [EMAIL PROTECTED] wrote in message
 [EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
  I have a doubt regarding adjusted r-square
 
  How do we calculate the adjusted r-square when the error degrees of
  freedom are zero ?
  (or in other words, number of samples is equal  to the number of
  regression terms including the constant)
 
  Such a situation leads to a zero in the denominator in the expression
  for calculating adjusted r-square.
 
  Your help is highly appreciated.
 
  Thanks
  Atul




 =
 Instructions for joining and leaving this list and remarks about
 the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
 =




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: adjusted r-square

2001-08-22 Thread Elliot Cramer

In sci.stat.consult Atul [EMAIL PROTECTED] wrote:
: I have a doubt regarding adjusted r-square

: How do we calculate the adjusted r-square when the error degrees of
: freedom are zero ?

You don't. you will have perfect prediction even for random numbers.


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: adjusted r-square

2001-08-22 Thread Jay Warner


I would like to say, 'on the contrary,' but I'm not contradicting Eric
and the others' comments RE: "you don't calculate r^2 with insufficient
data."
If the number of observations equals the number of terms in the (regression)
model you do have a perfect fit, with no df, etc. This data is, nonetheless,
information. To use the Baysian terminology, you must have prior
information (i.e., strong belief) that the statistical error (variance)
is smaller than some of the observed effects. Then you can use a
few techniques, such as probability plots or raw rank selection, to select
out the larger effects. Where possible, one can then go back, run
some confirmation trials and establish if these effects are 'real.'
Where confirmation trials are not possible the accepted approach
:) is to assert loudly which effects you believe in, and go with them.
You did the study to make a prediction and pick a path to give you an
advantage, yes? Believing the large effects is the way to bet.
What odds you give this path depends in part on how much data you have.
One qualifier to the above: I do most of my work with designed
experiments - near orthogonal arrays. If you have significant confounding
in your design, you need to be aware of what it is so that you are not
led down the traditional garden path by your own data.
Jay
Eric Bohlman wrote:
In sci.stat.consult Graeme Byrne [EMAIL PROTECTED]>
wrote:
> In short, you don't. If the number of terms in the model equals the
number
> of observations you have much bigger problems than not being able
to compute
> adjusted R^2. It should always be the case that the number of observations
> exceed the number of terms in the model otherwise you cannot calculate
any
> of the standard regression diagnostics (F-stats, t-stats etc). My
advice is
> get more data or remove terms from the model. If neither of these
is an
> option you are stuck.
It's worse than not being able to calculate regression diagnostics.
You
can't make *any* inferences beyond your observed data. Consider
the
degenerate case of trying to fit a bivariate regression line when you
have
only two observations. You'll *always* get a perfect fit because
two
points mathematically define a line. But that perfect fit will
tell you
absolutely nothing about the underlying relationship between the two
variables. It's consistent with *any* possible relationship,
including
complete independence. You can't tell how far off your model
is from the
observations because there simply isn't any room for it to be off ("room
for the model to be off" otherwise being known as "degrees of freedom").
A model with as many parameters as observations is equivalent to the
observations themselves, and therefore testing such a model against
the
observations is the same thing as asking if the observations are equal
to
themselves, which is circular reasoning.
=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at

http://jse.stat.ncsu.edu/
=

--
Jay Warner
Principal Scientist
Warner Consulting, Inc.
 North Green Bay Road
Racine, WI 53404-1216
USA
Ph: (262) 634-9100
FAX: (262) 681-1133
email: [EMAIL PROTECTED]
web: http://www.a2q.com
The A2Q Method (tm) -- What do you want to improve today?





=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=


adjusted r-square

2001-08-21 Thread Atul

I have a doubt regarding adjusted r-square

How do we calculate the adjusted r-square when the error degrees of
freedom are zero ?
(or in other words, number of samples is equal  to the number of
regression terms including the constant)

Such a situation leads to a zero in the denominator in the expression
for calculating adjusted r-square.

Your help is highly appreciated.

Thanks
Atul


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: adjusted r-square

2001-08-21 Thread Donald Burrill

On 21 Aug 2001, Atul wrote:

 How do we calculate the adjusted r-square when the error degrees of 
 freedom are zero ?  (Or in other words, number of samples is equal to 
 the number of regression terms including the constant.)
 Such a situation leads to a zero in the denominator in the expression
 for calculating adjusted r-square.

Depends in part on the expression you use, but in any case you also get 
a zero in the numerator.  Cf. Draper  Smith, Eq. (2.6.11b):  the 
right-hand expression indeed contains (n-p) in the denominator, but it 
also includes (1-R^2) in the numerator, which produces the indeterminate 
quotient 0/0.  In the middle expression of that equation, the quotient 
(residual SS)/(n-p) appears, which is also 0/0.

All of which only emphasizes that the result of any analysis for which 
the error d.f. = 0 is meaningless:  whether r-square, or regression 
coefficients, or error mean square, ... .  Statistical conclusions 
cannot, in general, be drawn from such an analysis.

 
 Donald F. Burrill [EMAIL PROTECTED]
 184 Nashua Road, Bedford, NH 03110  603-471-7128



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=