.. |
alpha_star
|
[RLlib] Async parallel execution manager. (#24423)
|
2022-05-25 17:54:08 +02:00 |
alpha_zero
|
[RLlib] AlphaZero TrainerConfig objects. (#25256)
|
2022-05-30 15:37:58 +02:00 |
ars
|
[RLlib] Upgrade gym 0.23 (#24171)
|
2022-05-23 08:18:44 +02:00 |
bandit
|
[RLlib] Agents to algos: DQN w/o Apex and R2D2, DDPG/TD3, SAC, SlateQ, QMIX, PG, Bandits (#24896)
|
2022-05-19 18:30:42 +02:00 |
cql
|
[RLlib]: Rename input_evaluation to off_policy_estimation_methods . (#25107)
|
2022-05-27 13:14:54 +02:00 |
ddpg
|
[RLlib] Async parallel execution manager. (#24423)
|
2022-05-25 17:54:08 +02:00 |
dqn
|
[RLlib] Migrate PPO Impala and APPO policies to use sub-classing implementation. (#25117)
|
2022-05-25 14:38:03 +02:00 |
dreamer
|
[rllib] Fix some missing agent->algorithm doc changes (#24841)
|
2022-05-16 11:52:49 +01:00 |
es
|
[RLlib] Upgrade gym 0.23 (#24171)
|
2022-05-23 08:18:44 +02:00 |
maddpg
|
[RLlib] MADDPG: Move into main algorithms folder and add proper unit and learning tests. (#24579)
|
2022-05-24 12:53:53 +02:00 |
maml
|
[RLlib] Migrate PPO Impala and APPO policies to use sub-classing implementation. (#25117)
|
2022-05-25 14:38:03 +02:00 |
marwil
|
[RLlib]: Rename input_evaluation to off_policy_estimation_methods . (#25107)
|
2022-05-27 13:14:54 +02:00 |
mbmpo
|
[RLlib] Migrate MAML, MB-MPO, MARWIL, and BC to use Policy sub-classing implementation. (#24914)
|
2022-05-20 14:10:59 +02:00 |
pg
|
[RLlib] APEX-DQN and R2D2 config objects. (#25067)
|
2022-05-23 12:15:45 +02:00 |
qmix
|
[RLlib] SAC, RNNSAC, and CQL TrainerConfig objects (#25059)
|
2022-05-22 19:58:47 +02:00 |
sac
|
[RLLib] Fix RNNSAC example failing on CI + fixes for recurrent models for other Q Learning Algos. (#24923)
|
2022-05-24 14:39:43 +02:00 |
slateq
|
[RLlib] SAC, RNNSAC, and CQL TrainerConfig objects (#25059)
|
2022-05-22 19:58:47 +02:00 |
__init__.py
|
[RLlib] Moved agents.es to algorithms.es (#24511)
|
2022-05-06 14:54:22 +02:00 |