ray/rllib/contrib
2021-11-23 23:01:05 +01:00
..
alpha_zero Revert "Revert [RLlib] POC: Deprecate build_policy (policy template) for torch only; PPOTorchPolicy (#20061) (#20399)" (#20417) 2021-11-16 14:49:41 +01:00
bandits [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
maddpg [RLlib] Replay buffer API (cleanups; docstrings; renames; move into rllib/execution/buffers dir) (#20552) 2021-11-19 11:57:37 +01:00
random_agent [RLlib] Trainer sub-class PPO/DDPPO (instead of build_trainer()). (#20571) 2021-11-23 23:01:05 +01:00
sumo [Lint] Add flake8-bugbear (#19053) 2021-10-03 23:24:11 -07:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
README.rst [docs] Move all /latest links to /master (#11897) 2020-11-10 10:53:28 -08:00
registry.py [RLlib] Trainer sub-class PPO/DDPPO (instead of build_trainer()). (#20571) 2021-11-23 23:01:05 +01:00

Contributed algorithms, which can be run via ``rllib train --run=contrib/<alg_name>``

See https://docs.ray.io/en/master/rllib-dev.html for guidelines.