PPO-PyTorch
PPO-PyTorch copied to clipboard
ValueError: expected sequence of length 8 at dim 1 (got 0)
Two lines should be changed in train.py state, _ = env.reset() # at 170 state, reward,done, _, _ = env.step(action) # at 177
Problem is solved. thx.