[ 
https://issues.apache.org/jira/browse/MAHOUT-228?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12794230#action_12794230
 ] 

Ted Dunning commented on MAHOUT-228:
------------------------------------


This implementation is purely logistic regression.  Changing to other 
supervised learning algorithms shouldn't be difficult and I have made the 
regularization pluggable, but I would as soon get this working as is before 
adding too much generality.  In particular, I have strongly used the 
presumption that I can do sparse updates and lazy regularization.  I don't know 
how much that applies to other problems.

> Need sequential logistic regression implementation using SGD techniques
> -----------------------------------------------------------------------
>
>                 Key: MAHOUT-228
>                 URL: https://issues.apache.org/jira/browse/MAHOUT-228
>             Project: Mahout
>          Issue Type: New Feature
>          Components: Classification
>            Reporter: Ted Dunning
>             Fix For: 0.3
>
>         Attachments: MAHOUT-228-1.patch
>
>
> Stochastic gradient descent (SGD) is often fast enough for highly scalable 
> learning (see Vowpal Wabbit, http://hunch.net/~vw/).
> I often need to have a logistic regression in Java as well, so that is a 
> reasonable place to start.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to