ray/rllib/agents/ppo
Sven Mika 499ad5fbe4
[RLlib] PyTorch version of APPO. (#8120)
- Translate all vtrace functionality to torch and added torch to the framework_iterator-loop in all existing vtrace test cases.
- Add learning test cases for APPO torch (both w/ and w/o v-trace).
- Add quick compilation tests for APPO (tf and torch, v-trace and no v-trace).
2020-04-23 09:11:12 +02:00
..
tests [RLlib] PyTorch version of APPO. (#8120) 2020-04-23 09:11:12 +02:00
__init__.py [rllib] Add Decentralized DDPPO trainer and documentation (#7088) 2020-02-10 15:28:27 -08:00
appo.py [RLlib] PyTorch version of APPO. (#8120) 2020-04-23 09:11:12 +02:00
appo_tf_policy.py [RLlib] PyTorch version of APPO. (#8120) 2020-04-23 09:11:12 +02:00
appo_torch_policy.py [RLlib] PyTorch version of APPO. (#8120) 2020-04-23 09:11:12 +02:00
ddppo.py [rllib] Rename sample_batch_size => rollout_fragment_length (#7503) 2020-03-14 12:05:04 -07:00
ppo.py [RLlib] Bug fix: Copy is_exploring placeholder for multi-GPU tower generation. (#7846) 2020-04-03 10:44:58 -07:00
ppo_tf_policy.py [RLlib] Exploration API: Policy changes needed for forward pass noisifications. (#7798) 2020-04-01 00:43:21 -07:00
ppo_torch_policy.py [RLlib] Exploration API: Policy changes needed for forward pass noisifications. (#7798) 2020-04-01 00:43:21 -07:00
utils.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00