I have the following actor-critic model


    class actor(gluon.HybridBlock):
        def __init__(self, r, **kwargs):
            super(actor, self).__init__(**kwargs)
            with self.name_scope():
                obvSize = 2*r + 1
                nOutputs = 2
                self.ac1 = gluon.nn.Dense(4, activation = 'relu')
                self.ac3 = gluon.nn.Dense(nOutputs)

        def hybrid_forward(self, f, x):
            ac2_i = self.ac1(x)
            #ac3_i = self.ac2(ac2_i)
            action = f.softmax(self.ac3(ac2_i))
            return action

    class critic(gluon.HybridBlock):
        def __init__(self, r, **kwargs):
            super(critic, self).__init__(**kwargs)
            with self.name_scope():
                obvSize = 2*r + 1
                nActions = 1
                self.cr1 = gluon.nn.Dense(4, activation = 'relu')
                self.cr3 = gluon.nn.Dense(1)

        def hybrid_forward(self, f, x):
            return self.cr3(self.cr1(x))

And I'm training the reinforcement learning agent over 5000episodes, and some 
of the weights of the actor becomes NA after 2000Episodes. (learning rate = 
0.001)

What can be the reason for that?





---
[Visit 
Topic](https://discuss.mxnet.io/t/weights-becomes-na-after-long-training/6511/1)
 or reply to this email to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.mxnet.io/email/unsubscribe/1f3badccc2145426a01fd905748b4016ae1e4b8fcc3ea146bf0a66a560074bfc).

Reply via email to