ray/rllib/agents/ppo
2020-11-26 13:14:11 +01:00
..
tests [RLlib] Torch LR schedule not working. Fix and added test case. (#12396) 2020-11-26 13:14:11 +01:00
__init__.py [RLlib] Examples folder restructuring (models) part 1 (#8353) 2020-05-08 08:20:18 +02:00
appo.py [RLlib] Fix inconsistency wrt batch size in SampleCollector (traj. view API). Makes DD-PPO work with traj. view API. (#12063) 2020-11-19 19:01:14 +01:00
appo_tf_policy.py [RLlib] Minor fixes (torch GPU bugs + some cleanup). (#11609) 2020-10-27 10:00:24 +01:00
appo_torch_policy.py [RLlib] Minor fixes (torch GPU bugs + some cleanup). (#11609) 2020-10-27 10:00:24 +01:00
ddppo.py [RLlib] Fix inconsistency wrt batch size in SampleCollector (traj. view API). Makes DD-PPO work with traj. view API. (#12063) 2020-11-19 19:01:14 +01:00
ppo.py [RLlib] Trajectory view API: Enable by default for PPO, IMPALA, PG, A3C (tf and torch). (#11747) 2020-11-12 16:27:34 +01:00
ppo_tf_policy.py [RLlib] Fix all example scripts to run on GPUs. (#11105) 2020-10-02 23:07:44 +02:00
ppo_torch_policy.py [RLlib] Trajectory view API: Enable by default for PPO, IMPALA, PG, A3C (tf and torch). (#11747) 2020-11-12 16:27:34 +01:00
README.md Use master for links to docs in source (#10866) 2020-09-19 00:30:45 -07:00

Proximal Policy Optimization (PPO)

Implementations of:

  1. Proximal Policy Optimization (PPO).

    Detailed Documentation

    Implementation

  2. Asynchronous Proximal Policy Optimization (APPO).

    Detailed Documentation

    Implementation

  3. Decentralized Distributed Proximal Policy Optimization (DDPPO)

    Detailed Documentation

    Implementation