ray/rllib/agents
2020-12-24 09:31:35 -05:00
..
a3c [RLlib] Batch-size for truncate_episode batch_mode should be confgurable in agent-steps (rather than env-steps), if needed. (#12420) 2020-12-08 16:41:45 -08:00
ars [RLlib] Fix most remaining RLlib algos for running with trajectory view API. (#12366) 2020-12-01 17:41:10 -08:00
ddpg [RLlib] Improved Documentation for PPO, DDPG, and SAC (#12943) 2020-12-24 09:31:35 -05:00
dqn [RLlib] Fix most remaining RLlib algos for running with trajectory view API. (#12366) 2020-12-01 17:41:10 -08:00
dreamer [rllib] Forgot to pass ioctx to child json readers (#11839) 2020-11-05 22:07:57 -08:00
es [RLlib] Fix most remaining RLlib algos for running with trajectory view API. (#12366) 2020-12-01 17:41:10 -08:00
impala Fix typo RMSProp -> RMSprop (#13063) 2020-12-23 13:37:46 -08:00
maml [RLLib] MAML extension for all models except RNNs (#11337) 2020-11-12 16:51:40 -08:00
marwil [RLlib] Batch-size for truncate_episode batch_mode should be confgurable in agent-steps (rather than env-steps), if needed. (#12420) 2020-12-08 16:41:45 -08:00
mbmpo MBMPO Cartpole (#11832) 2020-11-12 10:30:41 -08:00
pg [RLlib] Fix most remaining RLlib algos for running with trajectory view API. (#12366) 2020-12-01 17:41:10 -08:00
ppo [RLlib] Improved Documentation for PPO, DDPG, and SAC (#12943) 2020-12-24 09:31:35 -05:00
qmix [RLlib] Batch-size for truncate_episode batch_mode should be confgurable in agent-steps (rather than env-steps), if needed. (#12420) 2020-12-08 16:41:45 -08:00
sac [RLlib] Improved Documentation for PPO, DDPG, and SAC (#12943) 2020-12-24 09:31:35 -05:00
slateq [RLlib] Implement the SlateQ algorithm (#11450) 2020-11-03 09:52:04 +01:00
__init__.py [RLlib] MB-MPO cleanup (comments, docstrings, type annotations). (#11033) 2020-10-06 20:28:16 +02:00
callbacks.py [RLlib] Attention Nets: tf (#12753) 2020-12-20 20:22:32 -05:00
mock.py [tune] Use public methods for trainable (#9184) 2020-07-01 11:00:00 -07:00
registry.py [RLlib] Implement the SlateQ algorithm (#11450) 2020-11-03 09:52:04 +01:00
trainer.py [RLlib] eval_workers after restore not generated in Trainer due to unintuitive config handling. (#12844) 2020-12-20 09:37:31 -05:00
trainer_template.py WIP. (#12706) 2020-12-09 11:49:21 -08:00