ray/rllib/agents/ppo
2020-10-02 23:07:44 +02:00
..
tests [RLlib] Fix all example scripts to run on GPUs. (#11105) 2020-10-02 23:07:44 +02:00
__init__.py [RLlib] Examples folder restructuring (models) part 1 (#8353) 2020-05-08 08:20:18 +02:00
appo.py Enable vtrace by default (#10962) 2020-09-22 22:18:21 -07:00
appo_tf_policy.py [RLlib] Add docstrings for agents/dqn (#10710) 2020-09-15 12:37:07 +02:00
appo_torch_policy.py [RLlib] PPO, APPO, and DD-PPO code cleanup. (#10420) 2020-09-02 14:03:01 +02:00
ddppo.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
ppo.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
ppo_tf_policy.py [RLlib] Fix all example scripts to run on GPUs. (#11105) 2020-10-02 23:07:44 +02:00
ppo_torch_policy.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
README.md Use master for links to docs in source (#10866) 2020-09-19 00:30:45 -07:00

Proximal Policy Optimization (PPO)

Implementations of:

  1. Proximal Policy Optimization (PPO).

    Detailed Documentation

    Implementation

  2. Asynchronous Proximal Policy Optimization (APPO).

    Detailed Documentation

    Implementation

  3. Decentralized Distributed Proximal Policy Optimization (DDPPO)

    Detailed Documentation

    Implementation