On Tue, Dec 06, 2011 at 10:25:47PM +0100, Alexandre Gramfort wrote:
> regarding the scaling by n_samples using estimators I am convinced the right
> thing to do cf. my current PR to do this also on SVM models

I think that scaling by n_samples makes sense in the supervised learning
context (we often do the equivalent thing where we take the mean, rather than
the sum, over the unregularized training objective, making the regularization
invariant to the size of the training set), however there is a disconnect
between the dictionary learning notion of n_samples and the supervised
estimator notion of n_samples, and the conflation of these two because one
can be implemented by the other.

To be precise, (and I hope I got this right lest I confuse things further), a
sparse coding problem with K different training examples and L different
input features and M sparse components corresponds to K independent lasso
problems with L training examples each and M input features.

In this case, scaling the penalty by the Lasso "n_samples" corresponds to
scaling by the sparse coding "n_features", which I think you'll agree is a
bit weird.

David

------------------------------------------------------------------------------
Cloud Services Checklist: Pricing and Packaging Optimization
This white paper is intended to serve as a reference, checklist and point of 
discussion for anyone considering optimizing the pricing and packaging model 
of a cloud services business. Read Now!
http://www.accelacomm.com/jaw/sfnl/114/51491232/
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

Reply via email to