ray/rllib/contrib
Sven Mika 902e854af2
[RLlib; Docs overhaul] Docstring cleanup: Environments. (#19784)
* wip.

* Test: Make a change in tune to trigger tune tests, which are not run otherwise, but seem to fail nevertheless with this PR's changes.

* remove bare_metal_policy_with_custom_view_reqs from tests
2021-10-29 10:46:52 +02:00
..
alpha_zero [RLlib; Docs overhaul] Docstring cleanup: Environments. (#19784) 2021-10-29 10:46:52 +02:00
bandits [RLlib] Unify all RLlib Trainer.train() -> results[info][learner][policy ID][learner_stats] and add structure tests. (#18879) 2021-09-30 16:39:05 +02:00
maddpg [RLlib] Fix all the CI tests that were broken by is_training and replay buffer changes; re-comment-in the failing RLlib tests (#19809) 2021-10-28 18:06:47 +02:00
random_agent [tune] Use public methods for trainable (#9184) 2020-07-01 11:00:00 -07:00
sumo [Lint] Add flake8-bugbear (#19053) 2021-10-03 23:24:11 -07:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
README.rst [docs] Move all /latest links to /master (#11897) 2020-11-10 10:53:28 -08:00
registry.py [RLlib] Allow rllib rollout to run distributed via evaluation workers. (#13718) 2021-02-08 12:05:16 +01:00

Contributed algorithms, which can be run via ``rllib train --run=contrib/<alg_name>``

See https://docs.ray.io/en/master/rllib-dev.html for guidelines.