I have the following actor-critic model
class actor(gluon.HybridBlock):
def __init__(self, r, **kwargs):
super(actor, self).__init__(**kwargs)
with self.name_scope():
obvSize = 2*r + 1
nOutputs = 2
self.ac1 = gluon.nn.Dense(4, activation = 'relu')
self.ac3 = gluon.nn.Dense(nOutputs)
def hybrid_forward(self, f, x):
ac2_i = self.ac1(x)
#ac3_i = self.ac2(ac2_i)
action = f.softmax(self.ac3(ac2_i))
return action
class critic(gluon.HybridBlock):
def __init__(self, r, **kwargs):
super(critic, self).__init__(**kwargs)
with self.name_scope():
obvSize = 2*r + 1
nActions = 1
self.cr1 = gluon.nn.Dense(4, activation = 'relu')
self.cr3 = gluon.nn.Dense(1)
def hybrid_forward(self, f, x):
return self.cr3(self.cr1(x))
And I'm training the reinforcement learning agent over 5000episodes, and some
of the weights of the actor becomes NA after 2000Episodes. (learning rate =
0.001)
What can be the reason for that?
---
[Visit
Topic](https://discuss.mxnet.io/t/weights-becomes-na-after-long-training/6511/1)
or reply to this email to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.mxnet.io/email/unsubscribe/1f3badccc2145426a01fd905748b4016ae1e4b8fcc3ea146bf0a66a560074bfc).