ray/rllib/policy
2022-02-23 08:30:38 +01:00
..
tests [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
__init__.py [RLlib] JAXPolicy prep. PR #1. (#13077) 2020-12-26 20:14:18 -05:00
dynamic_tf_policy.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
eager_tf_policy.py Revert "Revert "[RLlib] Speedup A3C up to 3x (new training_iteration function instead of execution_plan) and re-instate Pong learning test."" (#18708) 2022-02-10 13:44:22 +01:00
policy.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
policy_map.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
policy_template.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
rnn_sequencing.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
sample_batch.py Revert "Revert "[RLlib] Speedup A3C up to 3x (new training_iteration function instead of execution_plan) and re-instate Pong learning test."" (#18708) 2022-02-10 13:44:22 +01:00
tf_policy.py Revert "Revert "[RLlib] AlphaStar: Parallelized, multi-agent/multi-GPU learni…" (#22153) 2022-02-08 16:43:00 +01:00
tf_policy_template.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
torch_policy.py [RLlib] Issue #21671: Handle callbacks and model metrics for TorchPolicy while using multi-GPU optimizers (#21697) 2022-02-23 08:30:38 +01:00
torch_policy_template.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
view_requirement.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00