ray/rllib/evaluation
2020-05-21 10:16:18 -07:00
..
__init__.py [rllib] Deprecate policy optimizers (#8345) 2020-05-21 10:16:18 -07:00
episode.py [RLlib] Remove TupleActions and support arbitrarily nested action spaces. (#8143) 2020-04-28 14:59:16 +02:00
metrics.py [rllib] Deprecate policy optimizers (#8345) 2020-05-21 10:16:18 -07:00
observation_function.py [rllib] observation function api for multi-agent (#8236) 2020-05-04 22:13:49 -07:00
policy_evaluator.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
policy_graph.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
postprocessing.py [rllib] implemented compute_advantages without gae (#6941) 2020-01-31 22:25:45 -08:00
rollout_metrics.py [rllib] Feature/histograms in tensorboard (#6942) 2020-01-30 22:02:53 -08:00
rollout_worker.py [rllib] Deprecate policy optimizers (#8345) 2020-05-21 10:16:18 -07:00
sample_batch.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
sample_batch_builder.py [RLlib] Fix for issue https://github.com/ray-project/ray/issues/8191 (#8200) 2020-04-27 23:19:26 +02:00
sampler.py [RLlib] Policy.compute_single_action() broken for nested actions (Issue 8411). (#8514) 2020-05-20 22:29:08 +02:00
tf_policy_graph.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
torch_policy_graph.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
worker_set.py [rllib] Deprecate policy optimizers (#8345) 2020-05-21 10:16:18 -07:00