ray/rllib/agents/ppo
2020-11-27 16:25:47 -08:00
..
tests [RLlib] Attention Net prep PR #1: Smaller cleanups. (#12447) 2020-11-27 16:25:47 -08:00
__init__.py [RLlib] Examples folder restructuring (models) part 1 (#8353) 2020-05-08 08:20:18 +02:00
appo.py [RLlib] Fix inconsistency wrt batch size in SampleCollector (traj. view API). Makes DD-PPO work with traj. view API. (#12063) 2020-11-19 19:01:14 +01:00
appo_tf_policy.py [RLlib] Minor fixes (torch GPU bugs + some cleanup). (#11609) 2020-10-27 10:00:24 +01:00
appo_torch_policy.py [RLlib] Attention Net prep PR #1: Smaller cleanups. (#12447) 2020-11-27 16:25:47 -08:00
ddppo.py [RLlib] Fix inconsistency wrt batch size in SampleCollector (traj. view API). Makes DD-PPO work with traj. view API. (#12063) 2020-11-19 19:01:14 +01:00
ppo.py [RLlib] Trajectory view API: Enable by default for PPO, IMPALA, PG, A3C (tf and torch). (#11747) 2020-11-12 16:27:34 +01:00
ppo_tf_policy.py [RLlib] Attention Net prep PR #1: Smaller cleanups. (#12447) 2020-11-27 16:25:47 -08:00
ppo_torch_policy.py [RLlib] Attention Net prep PR #1: Smaller cleanups. (#12447) 2020-11-27 16:25:47 -08:00
README.md Use master for links to docs in source (#10866) 2020-09-19 00:30:45 -07:00

Proximal Policy Optimization (PPO)

Implementations of:

  1. Proximal Policy Optimization (PPO).

    Detailed Documentation

    Implementation

  2. Asynchronous Proximal Policy Optimization (APPO).

    Detailed Documentation

    Implementation

  3. Decentralized Distributed Proximal Policy Optimization (DDPPO)

    Detailed Documentation

    Implementation