ray/rllib/agents/qmix
gjoliver 99a0088233
[RLlib] Unify the way we create local replay buffer for all agents (#19627)
* [RLlib] Unify the way we create and use LocalReplayBuffer for all the agents.

This change
1. Get rid of the try...except clause when we call execution_plan(),
   and get rid of the Deprecation warning as a result.
2. Fix the execution_plan() call in Trainer._try_recover() too.
3. Most importantly, makes it much easier to create and use different types
   of local replay buffers for all our agents.
   E.g., allow us to easily create a reservoir sampling replay buffer for
   APPO agent for Riot in the near future.
* Introduce explicit configuration for replay buffer types.
* Fix is_training key error.
* actually deprecate buffer_size field.
2021-10-26 20:56:02 +02:00
..
tests [RLlib] Issue 15973: Trainer.with_updates(validate_config=...) behaves confusingly. (#16429) 2021-06-19 22:42:00 +02:00
__init__.py [rllib] Deprecate policy optimizers (#8345) 2020-05-21 10:16:18 -07:00
mixers.py [RLlib] SAC algo cleanup. (#10825) 2020-09-20 11:27:02 +02:00
model.py [RLlib] Attention Net prep PR #3. (#12450) 2020-12-07 13:08:17 +01:00
qmix.py [RLlib] Unify the way we create local replay buffer for all agents (#19627) 2021-10-26 20:56:02 +02:00
qmix_policy.py [RLlib] Unify all RLlib Trainer.train() -> results[info][learner][policy ID][learner_stats] and add structure tests. (#18879) 2021-09-30 16:39:05 +02:00
README.md [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00

Code in this package is adapted from https://github.com/oxwhirl/pymarl.