[
https://issues.apache.org/jira/browse/SINGA-60?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14702987#comment-14702987
]
ASF subversion and git services commented on SINGA-60:
------------------------------------------------------
Commit 6afa895b8ea060a532ea01f1f4484c9db11a2496 in incubator-singa's branch
refs/heads/master from Wei Wang
[ https://git-wip-us.apache.org/repos/asf?p=incubator-singa.git;h=6afa895 ]
SINGA-60 Make learning rate and param init modular
Created a base class for getting learning rate, which is changed during
training.
Created a base class for initializing parameter values.
SINGA comes with a couple of built-in implementations for the two base
classes.
Users can also implement their own learning rate changing methods and
parameter initializing methods by extending the correponding base
classes.
> Make learning rate and param init modular
> -----------------------------------------
>
> Key: SINGA-60
> URL: https://issues.apache.org/jira/browse/SINGA-60
> Project: Singa
> Issue Type: Improvement
> Reporter: wangwei
>
> The learning rate of SGD typically changes through time.
> There are many different ways to change the learning rate of SGD. SINGA has
> implemented a couple of changing methods. But users may want to implement
> their own changing method. To make this part modular, this ticket is going to
> create a base learning rate generator, e.g. called LRGen, which is declared
> like,
> {code}
> class LRGenerator {
> public:
> virtual float Get (int step) = 0;
> protected:
> LRProto proto_;
> };
> {code}
> Users can then inherit LRGenerator to implement the their own changing
> algorithm in the `Get(int step)`.
> Users can also add configurations for their generator by extending the base
> LRProto.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)