Michal Figurski wrote:
Thank you all for your words of wisdom.

I start getting into what you mean by bootstrap. Not surprisingly, it seems to be something else than I do. The bootstrap is a tool, and I would rather compare it to a hammer than to a gun. People say that hammer is for driving nails. This situation is as if I planned to use it to break rocks.

The key point is that I don't really care about the bias or variance of the mean in the model. These things are useful for statisticians; regular people (like me, also a chemist) do not understand them and have no use for them (well, now I somewhat understand). My goal is very practical: I need an equation that can predict patient's outcome, based on some data, with maximum reliability and accuracy.

My two cents:

Bootstrapping (especially the optimism bootstrap, see Harrell 2001 ``Regression Modeling Strategies'') can be used to estimate how well a given model generalises. In other words, to estimate how much your model is overfitted to your data (more overfitting => less generalisable model).

This in itself is not useful for getting the coefficients of a good model (which is always done through MLE), but it can be used to compare different models. As Frank Harrell mentioned, you can do penalised regression, and find the best penalty through bootstrapping. This will possibly yield a model that is less overfitted and hence more reliable in terms of being valid for an unseen sample (from the same population).
Again, see Frank's book for more information about penalisation.

--
Gad Abraham
Dept. CSSE and NICTA
The University of Melbourne
Parkville 3010, Victoria, Australia
email: [EMAIL PROTECTED]
web: http://www.csse.unimelb.edu.au/~gabraham

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to