ray/rllib/agents/sac
Sven Mika 19c8033df2
[RLlib] Fix most remaining RLlib algos for running with trajectory view API. (#12366)
* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* WIP.

* LINT and fixes.
MB-MPO and MAML not working yet.

* wip

* update

* update

* rmeove

* remove dep

* higher

* Update requirements_rllib.txt

* Update requirements_rllib.txt

* relpos

* no mbmpo

Co-authored-by: Eric Liang <ekhliang@gmail.com>
2020-12-01 17:41:10 -08:00
..
tests [RLlib] Issue 11591: SAC loss does not use PR-weights in critic loss term. (#12394) 2020-11-25 11:28:46 -08:00
__init__.py [RLlib] SAC algo cleanup. (#10825) 2020-09-20 11:27:02 +02:00
README.md [docs] Move all /latest links to /master (#11897) 2020-11-10 10:53:28 -08:00
sac.py [RLlib] Fix most remaining RLlib algos for running with trajectory view API. (#12366) 2020-12-01 17:41:10 -08:00
sac_tf_model.py [RLlib] Support Simplex action spaces for SAC (torch and tf). (#11909) 2020-11-11 18:45:28 +01:00
sac_tf_policy.py [RLlib] Issue 11591: SAC loss does not use PR-weights in critic loss term. (#12394) 2020-11-25 11:28:46 -08:00
sac_torch_model.py [RLlib] Support Simplex action spaces for SAC (torch and tf). (#11909) 2020-11-11 18:45:28 +01:00
sac_torch_policy.py [RLlib] Attention Net prep PR #1: Smaller cleanups. (#12447) 2020-11-27 16:25:47 -08:00

Soft Actor Critic (SAC)

Implementations of:

Soft Actor-Critic Algorithm (SAC) and a discrete action extension.

Detailed Documentation

Implementation