unitree_rl_gym
unitree_rl_gym copied to clipboard
Why the policy is implemented using a multilayer perceptron (MLP) composed of one layers
class G1RoughCfgPPO( LeggedRobotCfgPPO ): class policy: init_noise_std = 0.8 actor_hidden_dims = [32] critic_hidden_dims = [32] activation = 'elu' # can be elu, relu, selu, crelu, lrelu, tanh, sigmoid # only for 'ActorCriticRecurrent': rnn_type = 'lstm' rnn_hidden_size = 64 rnn_num_layers = 1
I have the same question.
maybe just provide a sample example.