On Wed, Jan 8, 2020 at 9:38 PM lampahome <[email protected]> wrote:
> > > Stuart Reynolds <[email protected]> 於 2020年1月9日 週四 上午10:33寫道: > >> Correlated features typically have the property that they are tending to >> be similarly predictive of the outcome. >> >> L1 and L2 are both a preference for low coefficients. >> If a coefficient can be reduced yet another coefficient maintains similar >> loss, the these regularization methods prefer this solution. >> If you use L1 or L2, you should mean and variance normalize your features. >> >> > You mean LASSO and RIDGE both solve multilinearity? > LASSO has the reputation not to be good when there is multicollinearity, that's why elastic net L1 + L2 was introduced, AFAIK With multicollinearity the length of the parameter vector, beta' beta, is too large and L2, Ridge shrinks it. Josef > > _______________________________________________ > scikit-learn mailing list > [email protected] > https://mail.python.org/mailman/listinfo/scikit-learn >
_______________________________________________ scikit-learn mailing list [email protected] https://mail.python.org/mailman/listinfo/scikit-learn
