ray/rllib/agents
2020-03-04 10:03:41 -08:00
..
a3c [rllib] Support multi-agent training in pipeline impls, add easy flag to enable (#7338) 2020-03-02 15:16:37 -08:00
ars [RLlib] Add torch flag to train.py (#6807) 2020-01-17 18:48:44 -08:00
ddpg Checkpoint Adam momenta for DDPG (#7449) 2020-03-04 10:03:41 -08:00
dqn [RLlib] DDPG refactor and Exploration API action noise classes. (#7314) 2020-03-01 11:53:35 -08:00
es [RLlib] Add torch flag to train.py (#6807) 2020-01-17 18:48:44 -08:00
impala [RLlib] Exploration API: merge deterministic flag with exploration classes (SoftQ and StochasticSampling). (#7155) 2020-02-19 12:18:45 -08:00
marwil [rllib] implemented compute_advantages without gae (#6941) 2020-01-31 22:25:45 -08:00
pg [rllib] Support multi-agent training in pipeline impls, add easy flag to enable (#7338) 2020-03-02 15:16:37 -08:00
ppo Fix issue with torch PPO not handling action spaces of shape=(>1,). (#7398) 2020-03-02 10:53:19 -08:00
qmix [RLlib] Policy.compute_log_likelihoods() and SAC refactor. (issue #7107) (#7124) 2020-02-22 14:19:49 -08:00
sac Fix SAC bug (twin Q not used for min'ing over both Q-nets in loss func). (#7354) 2020-02-27 12:49:08 -08:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
agent.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
mock.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
registry.py [rllib] Support multi-agent training in pipeline impls, add easy flag to enable (#7338) 2020-03-02 15:16:37 -08:00
trainer.py [RLlib] Make rollout always use evaluation_config. (#7396) 2020-03-03 17:20:35 -08:00
trainer_template.py [rllib] Support multi-agent training in pipeline impls, add easy flag to enable (#7338) 2020-03-02 15:16:37 -08:00