ray/rllib/policy
Raphael Avalos 440c9c42be
[RLlib] Fix combination of lockstep and multiple agnts controlled by the same policy. (#9521)
* Change aggregation when lockstep is activated.

Modification of MultiAgentBatch.timeslices to support the combination of lockstep and multiple agents controlled by the same policy.

fix ray-project/ray#9295

* Line too long.
2020-07-19 23:03:12 -07:00
..
tests [RLlib] Tf2.x native. (#8752) 2020-07-11 22:06:35 +02:00
__init__.py [rllib] Add type annotations for evaluation/, env/ packages (#9003) 2020-06-19 13:09:05 -07:00
dynamic_tf_policy.py [RLlib] Type annotations for policy. (#9248) 2020-07-05 13:09:51 +02:00
eager_tf_policy.py [RLlib] Tf2.x native. (#8752) 2020-07-11 22:06:35 +02:00
policy.py [RLlib] Layout of Trajectory View API (new class: Trajectory; not used yet). (#9269) 2020-07-14 04:27:49 +02:00
rnn_sequencing.py [RLlib] Type annotations for policy. (#9248) 2020-07-05 13:09:51 +02:00
sample_batch.py [RLlib] Fix combination of lockstep and multiple agnts controlled by the same policy. (#9521) 2020-07-19 23:03:12 -07:00
tf_policy.py [RLlib] Tf2.x native. (#8752) 2020-07-11 22:06:35 +02:00
tf_policy_template.py [RLlib] Type annotations for policy. (#9248) 2020-07-05 13:09:51 +02:00
torch_policy.py [RLlib] Issue 9218: PyTorch Policy places Model on GPU even with num_gpus=0 (#9516) 2020-07-17 05:53:25 +02:00
torch_policy_template.py [RLlib] Minor cleanup in preparation to tf2.x support. (#9130) 2020-06-25 19:01:32 +02:00
trajectory_view.py [RLlib] Remove requirement for dataclasses in rllib (not supported in py3.5) (#9237) 2020-07-01 17:31:44 +02:00