hindsight-experience-replay icon indicating copy to clipboard operation
hindsight-experience-replay copied to clipboard

what is the definition of actor_loss in ddpg_agent,.py?

Open whynpt opened this issue 3 years ago • 1 comments

 actor_loss = -self.critic_network(inputs_norm_tensor, actions_real).mean()
 actor_loss += self.args.action_l2 * (actions_real / self.env_params['action_max']).pow(2).mean()

I think the output of critic_network is enough to be the actor_loss. So is it a regularizer or trick? it would be better for me to reply in Chinese.

whynpt avatar Dec 20 '21 12:12 whynpt

@whynpt It's more like a regularizer, make sure the action will not move too much.

TianhongDai avatar Dec 25 '21 01:12 TianhongDai