mirror of
https://github.com/vale981/ray
synced 2025-03-06 18:41:40 -05:00
24 lines
883 B
Markdown
24 lines
883 B
Markdown
![]() |
Proximal Policy Optimization (PPO)
|
||
|
==================================
|
||
|
|
||
|
Implementations of:
|
||
|
|
||
|
1) Proximal Policy Optimization (PPO).
|
||
|
|
||
|
**[Detailed Documentation](https://docs.ray.io/en/latest/rllib-algorithms.html#ppo)**
|
||
|
|
||
|
**[Implementation](https://github.com/ray-project/ray/blob/master/rllib/agents/ppo/ppo.py)**
|
||
|
|
||
|
2) Asynchronous Proximal Policy Optimization (APPO).
|
||
|
|
||
|
**[Detailed Documentation](https://docs.ray.io/en/latest/rllib-algorithms.html#appo)**
|
||
|
|
||
|
**[Implementation](https://github.com/ray-project/ray/blob/master/rllib/agents/ppo/appo.py)**
|
||
|
|
||
|
3) Decentralized Distributed Proximal Policy Optimization (DDPPO)
|
||
|
|
||
|
**[Detailed Documentation](https://docs.ray.io/en/latest/rllib-algorithms.html#decentralized-distributed-proximal-policy-optimization-dd-ppo)**
|
||
|
|
||
|
**[Implementation](https://github.com/ray-project/ray/blob/master/rllib/agents/ppo/ddppo.py)**
|
||
|
|