ray/rllib/agents
2020-04-30 11:09:33 -07:00
..
a3c [RLlib] SAC Torch (incl. Atari learning) (#7984) 2020-04-15 13:25:16 +02:00
ars [RLlib] Deprecate all Model(v1) usage. (#8146) 2020-04-29 12:12:59 +02:00
ddpg [RLlib] Deprecate all Model(v1) usage. (#8146) 2020-04-29 12:12:59 +02:00
dqn [rllib] Port DQN/Ape-X to training workflow api (#8077) 2020-04-23 12:39:19 -07:00
es [RLlib] Deprecate all Model(v1) usage. (#8146) 2020-04-29 12:12:59 +02:00
impala [RLlib] PyTorch version of APPO. (#8120) 2020-04-23 09:11:12 +02:00
marwil [RLlib] MARWIL torch. (#7836) 2020-04-06 16:38:50 -07:00
pg [RLlib] SAC Torch (incl. Atari learning) (#7984) 2020-04-15 13:25:16 +02:00
ppo [rllib] Execute PPO using training workflow (#8206) 2020-04-30 01:18:09 -07:00
qmix [RLlib] Deprecate all Model(v1) usage. (#8146) 2020-04-29 12:12:59 +02:00
sac [RLlib] Beta distribution. (#8229) 2020-04-30 11:09:33 -07:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
agent.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
callbacks.py [RLlib] Added DefaultCallbacks which replaces old callbacks dict interface (#6972) 2020-04-16 16:06:42 -07:00
mock.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
registry.py [rllib] Support multi-agent training in pipeline impls, add easy flag to enable (#7338) 2020-03-02 15:16:37 -08:00
trainer.py [rllib] Execute PPO using training workflow (#8206) 2020-04-30 01:18:09 -07:00
trainer_template.py [rllib] Execute PPO using training workflow (#8206) 2020-04-30 01:18:09 -07:00