Hi,
I run the following tuning function for svm. It's very strange that every
time i run this function, the best.parameters give different values.
[A]
svm.tune - tune(svm, train.x, train.y,
validation.x=train.x, validation.y=train.y,
ranges = list(gamma =
Maggie Wang wrote:
Hi,
I run the following tuning function for svm. It's very strange that every
time i run this function, the best.parameters give different values.
[A]
svm.tune - tune(svm, train.x, train.y,
validation.x=train.x, validation.y=train.y,
Thank you so much! I will have a try!! ~ maggie
On Dec 27, 2007 6:43 PM, Uwe Ligges [EMAIL PROTECTED]
wrote:
Maggie Wang wrote:
Hi, Uwe,
Thanks for the reply!! I have 87 observations in total. If this amount
causes the different best.parameters, is there a better way than cross
Maggie Wang wrote:
Hi, Uwe,
Thanks for the reply!! I have 87 observations in total. If this amount
causes the different best.parameters, is there a better way than cross
validation to tune them?
In order to get stable (I do not say best) results, you could try some
bootstrap with many
Hi, Uwe,
Thanks for the reply!! I have 87 observations in total. If this amount
causes the different best.parameters, is there a better way than cross
validation to tune them?
Thank you so much for the help!
Best Regards,
Maggie
On Dec 27, 2007 6:17 PM, Uwe Ligges [EMAIL PROTECTED]
wrote:
5 matches
Mail list logo