Sven Mika
|
130b7eeaba
|
[RLlib] Trainer to Algorithm renaming. (#25539)
|
2022-06-11 15:10:39 +02:00 |
|
Sven Mika
|
b5bc2b93c3
|
[RLlib] Move all remaining algos into algorithms directory. (#25366)
|
2022-06-04 07:35:24 +02:00 |
|
Yi Cheng
|
fd0f967d2e
|
Revert "[RLlib] Move (A/DD)?PPO and IMPALA algos to algorithms dir and rename policy and trainer classes. (#25346)" (#25420)
This reverts commit e4ceae19ef .
Reverts #25346
linux://python/ray/tests:test_client_library_integration never fail before this PR.
In the CI of the reverted PR, it also fails (https://buildkite.com/ray-project/ray-builders-pr/builds/34079#01812442-c541-4145-af22-2a012655c128). So high likely it's because of this PR.
And test output failure seems related as well (https://buildkite.com/ray-project/ray-builders-branch/builds/7923#018125c2-4812-4ead-a42f-7fddb344105b)
|
2022-06-02 20:38:44 -07:00 |
|
Sven Mika
|
e4ceae19ef
|
[RLlib] Move (A/DD)?PPO and IMPALA algos to algorithms dir and rename policy and trainer classes. (#25346)
|
2022-06-02 16:47:05 +02:00 |
|
Jun Gong
|
eaf9c941ae
|
[RLlib] Migrate PPO Impala and APPO policies to use sub-classing implementation. (#25117)
|
2022-05-25 14:38:03 +02:00 |
|
Sven Mika
|
44773e810b
|
[RLlib] DD-PPO Config objects. (#25028)
|
2022-05-22 13:05:24 +02:00 |
|
Sven Mika
|
026849cd27
|
[RLlib] APPO TrainerConfig objects. (#24376)
|
2022-05-02 15:06:23 +02:00 |
|
Sven Mika
|
2eaa54bd76
|
[RLlib] POC: Config objects instead of dicts (PPO only). (#23491)
|
2022-03-31 18:26:12 +02:00 |
|
Sven Mika
|
5f278c6411
|
[RLlib] Examples folder restructuring (models) part 1 (#8353)
|
2020-05-08 08:20:18 +02:00 |
|
Eric Liang
|
026f6884b5
|
[rllib] Add Decentralized DDPPO trainer and documentation (#7088)
|
2020-02-10 15:28:27 -08:00 |
|
Sven Mika
|
c957ed58ed
|
[RLlib] Implement PPO torch version. (#6826)
|
2020-01-20 23:06:50 -08:00 |
|
Eric Liang
|
5d7afe8092
|
[rllib] Try moving RLlib to top level dir (#5324)
|
2019-08-05 23:25:49 -07:00 |
|