I believe we need to know the following about packages:
(1) Does the package do what it purports to do, i.e. are the results valid?
(2) Have the results generated by the package been validate against some other 
statistical package, or hand-worked example?
(3) Are the methods used in the soundly based?
(4) Does the package documentation refer to referred papers or textbooks?
(5) In addition to the principle result, does the package return ancillary 
values that allow for proper interpretation of the main result, (e.g. lm gives 
estimates of the betas and their SEs, but also generates residuals)?.
(6) Is the package easy to use, i.e. do the parameters used when invoking the 
package chosen so as to allow the package to be flexible?
(7) Are the error messages produced by the package helpful?
(8) Does the package conform to standards of R coding and good programming 
principles in general?
(9) Does the package interact will with the larger R environment, e.g. does it 
have a plot method etc.?
(10) Is the package well documented internally, i.e. is the code easy to 
follow, are the comments in the code adequate?
(11) Is the package well documented externally, i.e. through man pages and 
perhaps other documentation (e.g. MASS and its associated textbook)?

In addition to package evaluation and reviews, we also need some plan for the 
future of R. Who will maintain, modify, and extend packages after the principle 
author, or authors, retire? Software is never "done". Errors need to be 
corrected, programs need to be modified to accommodate changes in software and 
hardware. I have reasonable certainty that commercial software (e.g. SAS) will 
be available in 10-years (and that PROC MIXED will still be a part of SAS). I 
am far less sanguine about any number of R packages.
John 



John Sorkin M.D., Ph.D.
Chief, Biostatistics and Informatics
University of Maryland School of Medicine Division of Gerontology
Baltimore VA Medical Center
10 North Greene Street
GRECC (BT/18/GR)
Baltimore, MD 21201-1524
(Phone) 410-605-7119
(Fax) 410-605-7913 (Please call phone number above prior to faxing)

>>> <[EMAIL PROTECTED]> 12/1/2007 2:21 AM >>>
This seems a little impractical to me.  People respond so much at random
and most only tackle questions with which they feel comfortable.  As
it's not a competition in any sense, it's going to be hard to rank
people in any effective way.  But suppose you succeed in doing so, then
what?

To me a much more urgent initiative is some kind of user online review
system for packages, even something as simple as that used by Amazon.com
has for customer review of books.

I think the need for this is rather urgent, in fact.  Most packages are
very good, but I regret to say some are pretty inefficient and others
downright dangerous.  You don't want to discourage people from
submitting their work to CRAN, but at the same time you do want some
mechanism that allows users to relate their experience with it, good or
bad.   


Bill Venables
CSIRO Laboratories
PO Box 120, Cleveland, 4163
AUSTRALIA
Office Phone (email preferred): +61 7 3826 7251
Fax (if absolutely necessary):  +61 7 3826 7304
Mobile:                         +61 4 8819 4402
Home Phone:                     +61 7 3286 7700
mailto:[EMAIL PROTECTED] 
http://www.cmis.csiro.au/bill.venables/ 

-----Original Message-----
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] 
On Behalf Of Doran, Harold
Sent: Saturday, 1 December 2007 6:13 AM
To: R Help
Subject: [R] Rating R Helpers

Since R is open source and help may come from varied levels of
experience on R-Help, I wonder if it might be helpful to construct a
method that can be used to "rate" those who provide help on this list.

This is something that is done on other comp lists, like
http://www.experts-exchange.com/.

I think some of the reasons for this are pretty transparent, but I
suppose one reason is that one could decide to implement the advise of
those with "superior" or "expert" levels. In other words, you can trust
the advice of someone who is more experienced more than someone who is
not. Currently, there is no way to discern who on this list is really an
R expert and who is not. Of course, there is R core, but most people
don't actually know who these people are (at least I surmise that to be
true).

If this is potentially useful, maybe one way to begin the development of
such ratings is to allow the original poster to "rate" the level of help
from those who responded. Maybe something like a very simple
questionnaire on a likert-like scale that the original poster would
respond to upon receiving help which would lead to the accumulation of
points for the responders. Higher points would result in higher levels
of expertise (e.g., novice, ..., wizaRd).

Just a random thought. What do others think?

Harold




        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help 
PLEASE do read the posting guide
http://www.R-project.org/posting-guide.html 
and provide commented, minimal, self-contained, reproducible code.

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help 
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html 
and provide commented, minimal, self-contained, reproducible code.

Confidentiality Statement:
This email message, including any attachments, is for th...{{dropped:6}}

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to