adventures-in-ml-code icon indicating copy to clipboard operation
adventures-in-ml-code copied to clipboard

This repository holds all the code for the site http://www.adventuresinmachinelearning.com

Results 26 adventures-in-ml-code issues
Sort by recently updated
recently updated
newest added

Hi, there is a bg in `policy_gradient_reinforce_tf2.py` at `line 39`. `loss = network.train_on_batch(states, discounted_rewards)` to fix this I made two changes, 1. one_hot_encode the actions `one_hot_encode = np.array([[1 if a==i...

I am getting this error while running the code given in file "per_duelingq_spaceinv_tf2.py" on Google Colab that uses the Tensorflow version: 2.3.0 ``` ValueError Traceback (most recent call last) in...

Hi. The code [Code](https://github.com/adventuresinML/adventures-in-ml-code/blob/master/policy_gradient_reinforce_tf2.py ) is not working with this line: `loss = network.train_on_batch(states, discounted_rewards)`.

loss = update_network(network, rewards, states, actions, num_actions) loss = network.train_on_batch(states, discounted_rewards)

First of all, thank you for the tutorial [here](https://adventuresinmachinelearning.com/policy-gradient-tensorflow-2/)! I am trying to implement/run your code mentioned in the tutorial, however, the results are not converging after 500 steps as...

``` # # start in commandline: python keras_lstm.py [-h] [--data_path DATA PATH] runopt # 'An integer: 1 to train, 2 to test' # i.e.: python keras_lstm.py 1 # or: python...

``` """ This is the jupyter notebook version of the tutorial with some small fixes. Instead of running it in command line like "python keras_lstm.py 1" with runopt parameter here...

ValueError Traceback (most recent call last) in () 56 57 if done: ---> 58 loss = update_network(network, rewards, states, actions, num_actions) 59 tot_reward = sum(rewards) 60 print(f"Episode: {episode}, Reward: {tot_reward},...

Should not the model be fitted (keras.fit(...)) and predicted (keras.predict(state)) in double Q learning (and also in dueling Q learning) examples? Seems you also forget to apply the same in...

Can you please provide a tutorial on how to run tf_word2vec.py in hadoop , basically to distribute the workload and then reduce.