Inverse-Reinforcement-Learning
Inverse-Reinforcement-Learning copied to clipboard
Implementations of selected inverse reinforcement learning algorithms.
Hi Matthew! This repo is just great: It works, its transparant and modular! I only found two differences between Ziebart's thesis and your implementation. Can you let me know if...
Hello!I am a graduate student from China,and in your demo,the feature_matrix is not provided,so when giving state space,how to obtain the feature_matrix?I would be grateful if you could answer. Apart...
I am currently running code and am getting an error: I reshaped the feature matrix as reshaped_to_2d_reshape as (num_of_states, num_of_dimensions) and am still getting this error. I am not sure...
I need help to try to remove this warning: WARNING (theano.configdefaults): g++ not available, if using conda: `conda install m2w64-toolchain` C:\Users\Sankalp Chauhan\AppData\Local\Programs\Python\Python37\lib\site-packages\theano\configdefaults.py:560: UserWarning: DeprecationWarning: there is no c++ compiler.This is...
I would like to ask you if you can list some references in order to understand how you formulated the block matrix form of the linear program for solving large...
Can you give some points on how you designed the feature matrix? You kept it as 25*25 (in case of Gridword), where each state is represented separately. According to my...
I was wondering why in irl/maxnet.py line 71, you unpack 3 thing from a 2-d array, I think this is a small error here?
In the backwards pass of MaxEnt (Algo 9.1 Brian's thesis), MaxEnt uses a softmax calculation to update the `V` function (soft Value function), but maxent.py seems to call value_iteration.optimal_value which...
The environments of GridWorld and ObjectWorld are all tabular environments, in which the states are discreate and limited. We can easily write down the feature matrix by listing all possible...