Hello,
I'm dealing with a problem without much data. As a solution, I'm training
10 estimators using a 10-Fold CV-Schema. Now, I wanted to persist those
models. In order to avoid having to save 10 estimators, I was thinking
about saving a single VotingRegressor with those pre-trained models or
Hi Samir, the following visualization might be useful for gaining intuition
on the meaning of a negative r2:
https://gist.github.com/WittmannF/02060b45ce3ec9239898a5b91df2564e
A negative r2 is reflects into a model predicting the opposite trend of the
data.
On Sat, Aug 14, 2021, 03:17 Samir K
In my opinion the reference is distorting a concept that has a consolidated
definition in the community. I am also familiar with the definition of WL
as "an estimator slightly better than guessing", mostly decision stumps (
https://en.m.wikipedia.org/wiki/Decision_stump), which is not an component
Hello guys,
The the following reference states that Random Forests uses weak learners:
-
ment would
be repeated in a given tree). My apologies. Everything makes sense again
On Sun, May 10, 2020, 19:42 Fernando Marcos Wittmann <
fernando.wittm...@gmail.com> wrote:
> Okay, so it's sampling with replacement with same size of the original
> dataset. That mean that som
My question is why the full dataset is being used as default when building
each tree. That's not random forest. The main point of RF is to build each
tree with a subsample of the full dataset
On Sun, May 10, 2020, 09:50 Joel Nothman wrote:
> A bootstrap is very commonly a random draw with
When reading the documentation of Random Forest, I got the following:
```
max_samples : int or float, default=None If bootstrap is True, the number
of samples to draw from X to train each base estimator. - *If None
(default), then draw `X.shape[0]` samples.* - If int, then draw
`max_samples`
That's an excellent discussion! I've always wondered how other tools like R
handled naturally categorical variables or not. LightGBM has a scikit-learn
API which handles categorical features by inputting their columns names (or
indexes):
```
import lightgbm
lgb=lightgbm.LGBMClassifier()
What about converting into two columns? One with the real projection and
the other with the complex projection?
On Sat, Oct 19, 2019, 3:44 PM ahmad qassemi wrote:
> Dear Mr/Mrs,
>
> I'm a PhD student in DS. I'm trying to use your provided code on *Spectral
> CoClustering *and *Spectral
Hi Sarah, I have some reflection questions. You don't need to answer all
of them :) how many categories (approximately) do you have in each of those
20M categorical variables? How many samples do you have? Maybe you should
consider different encoding strategies such as binary encoding. Also, this
>> scikit-learn mailing list
>> scikit-learn@python.org
>> https://mail.python.org/mailman/listinfo/scikit-learn
>>
>
> ___
> scikit-learn mailing list
> scikit-learn@python.org
_() got multiple values for keyword argument 'n_splits'
>
>
>
>
> --
> *Thanks,*
> *Shubham Singh Tomar*
> *Autodidact24.github.io <http://Autodidact24.github.io>*
>
> ___
> scikit-learn
12 matches
Mail list logo