agents icon indicating copy to clipboard operation
agents copied to clipboard

TF-Agents: A reliable, scalable and easy to use TensorFlow library for Contextual Bandits and Reinforcement Learning.

Results 191 agents issues
Sort by recently updated
recently updated
newest added

​import random from abc import ABC from random import choice import numpy as np from tf_agents.environments import py_environment from tf_agents.specs import is_continuous from tf_agents.trajectories import time_step as ts, time_step class...

Because tf.keras.utils.plot_model does not work with the agent network so I made a custom function to do this Example: ![](https://i.imgur.com/JosTHl5.jpg)

Hi there, I'd like to follow the [TF-Agents Environments Tutorial](https://www.tensorflow.org/agents/tutorials/2_environments_tutorial). However, I'm encountering an error that says "No module named 'tensorflow.python.training.tracking'" on this line -> from tf_agents.environments import py_environment. I'm...

Hi Team, thanks for creating this wonderful library. Can you guide where I can find a Multi-arm Multi-play Setting example? Note: I am familiar with Multi-arm Single-play setting example of...

Hi Team, Thanks for creating such wonderful notebooks and examples. 😊 I have successfully trained a model(For movie lens data) using a training script mentioned in the example section(https://github.com/tensorflow/agents/blob/master/tf_agents/bandits/agents/examples/v2/train_eval_movielens.py) Can...

I have a Keras model and I want to use it as the q network for DDQN. I don't know how to do it. ```python def model(): x = Input(shape=(_pkt_num,...

It seems to be necessary to increase the buffer_size for AverageReturnMetric to get better precision. One would think that increasing buffer_size would only speed up the computation (at a cost...

Currently, DQNAgent updates the target network periodically. Would it make sense to provide a target_update method in DQNAgent to update the target network? Reasons: 1. I find it confusing to...

import tensorflow as tf from tf_agents.environments import suite_gym from tf_agents.environments import tf_py_environment from tf_agents.environments import parallel_py_environment # 创建单个CartPole环境 def get_CartPole_Env(): return suite_gym.load("CartPole-v0") # 创建并行环境 tf_env = tf_py_environment.TFPyEnvironment( parallel_py_environment.ParallelPyEnvironment([get_CartPole_Env] * 4)...

Hello everyone, I wanted to run this code on Windows, Unfortunately reverb is only available for Linux. Do you have a functioning version of this Code for windows? Or do...