ElegantRL icon indicating copy to clipboard operation
ElegantRL copied to clipboard

Massively Parallel Deep Reinforcement Learning. 🔥

Results 156 ElegantRL issues
Sort by recently updated
recently updated
newest added

执行 `python demo_A2C_PPO.py --gpu=0 --drl=0 --env=6` 出现异常 ``` File "elegantrl/train/evaluator.py", line 176, in get_cumulative_rewards_and_steps tensor_action = tensor_action.argmax(dim=1) IndexError: Dimension out of range (expected to be in range of [-1, 0],...

bug

本人的环境是这样的: 操作系统:win10 cuda:11.8 cudnn:8.8.1 python:3.9.13 pytorch:2.0.0 ElegantRL是最新开发版 情况描述: 如果使用train_agent_multiprocessing,Learner进程通过管道把actor发到work进程这里: ``` '''Learner send actor to Workers''' for send_pipe in self.send_pipes: send_pipe.send(agent.act) ``` agent.act的成员state_std,会从 tensor([1., 1., 1., 1., 1., 1., 1., 1.,...

bug

您好,我直接使用demo_A2C_PPO.py训练pendulum环境下的A2C算法无法收敛,可能算法实现上有问题。AgentDiscreteA2C算法仅继承了AgentDiscretePPO,并未实现自己的update_net函数

bug

第235行部分self.cri.state_std[:] = self.cri.state_std,这样的话cri.state_std是永远不变的

bug

`AgentMATD3` & `AgentMADDPG` cannot be call by `train_agent` or `train_agent_multiprocessing`, when I add them to 'demo_DDPG_TD3_SAC.py' Error shows like that Traceback (most recent call last): File "/Users/c/Downloads/ElegantRL-master/examples/demo_DDPG_TD3_SAC.py", line 238, in...

bug

Are there any demos that can illustrate the process of using MATD3 / MADDPG to process discrete actions?

bug

---> 5 from elegantrl.train.run import Arguments 6 from numpy import random as rd 7 ImportError: cannot import name 'Arguments' from 'elegantrl.train.run' (/usr/local/lib/python3.8/dist-packages/elegantrl/train/run.py)

bug

在知乎上看到可以开启multiworkers进行探索,但是不知道在哪里可以开启这个功能?

bug

I'm confused why we use 'logprob = dist.log_prob(a_avg)' instead of 'logprob = dist.log_prob(action)' in line 247 of elegantrl/agents/net.py. I think the latter is consistent to the original paper. Is using...

dicussion

![image](https://user-images.githubusercontent.com/8471303/221737009-e6bacee7-7d0d-45ff-9388-f431cd1c5f90.png) in "demo_mujoco_draw_obj_h.py" ![image](https://user-images.githubusercontent.com/8471303/221737030-3c587e27-0d08-446e-96d8-ac351b1c6281.png) in "demo_Isaac_Gym.py" ![image](https://user-images.githubusercontent.com/8471303/221737081-819490bf-7d5b-4219-847b-a44bf009b7a8.png) in "demo_DDPG_H.py" ![image](https://user-images.githubusercontent.com/8471303/221737100-95a1bb2b-a1db-4a01-8d19-7cce3d72ec0c.png) in "demo_vec_env_A2C_PPO.py" ![image](https://user-images.githubusercontent.com/8471303/221737126-5feed013-f454-403c-987c-b46eaf9dfff4.png) in "demo_PPO_H.py" ![image](https://user-images.githubusercontent.com/8471303/221737163-17e6d5f4-b984-4b2f-88a6-994e79c1648c.png) in "demo_PaperTradingEnv_PPO.py" ![image](https://user-images.githubusercontent.com/8471303/221737186-f36fc2f9-9474-427a-8425-b6792ceae8bb.png) in "demo_mujoco_render.py" ![image](https://user-images.githubusercontent.com/8471303/221737215-f64ca0f6-89c1-43d5-91af-d1381cddb4c6.png) in "tutorial_LunarLanderContinous-v2.py" ![image](https://user-images.githubusercontent.com/8471303/221737244-3af17f8e-b11a-4e6c-b1c8-26b8553c81f0.png) in "tutorial_Hopper-v3.py" ![image](https://user-images.githubusercontent.com/8471303/221737275-659c8710-eb17-4732-948d-5645f122fd9f.png) in "...

bug