[
https://issues.apache.org/jira/browse/MAHOUT-702?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13035823#comment-13035823
]
Ted Dunning commented on MAHOUT-702:
------------------------------------
Nice work Hector.
I have a few comments.
First, doesn't your train method destroy the input vector? That seems like bad
manners. I think that you can get the
effect you want without an additional copy being made by accumulating into the
two rows of the weights matrix.
Secondly, I see why you put the test into the existing test so that you could
re-use some framework.
My preference is to keep a bit of separation, however. What do you think about
factoring out the
common structure and having both kinds of test extend the same abstract class?
Also, does your PA learner have any regularization other than early stopping?
What about annealing
of the learning rate?
Finally, what do you think about putting this under a similar framework as
AdaptiveLogisticRegression
in order to get auto-tuning of the learning rate?
> Implement Online Passive Aggressive learner
> -------------------------------------------
>
> Key: MAHOUT-702
> URL: https://issues.apache.org/jira/browse/MAHOUT-702
> Project: Mahout
> Issue Type: New Feature
> Components: Classification
> Affects Versions: 0.6
> Reporter: Hector Yee
> Priority: Minor
> Attachments: MAHOUT-702.patch
>
> Original Estimate: 24h
> Remaining Estimate: 24h
>
> Implements online passive aggressive learner that minimizes label ranking
> loss.
--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira