ray/rllib/agents/sac
maxco2 b8436f0f00
[rllib] Fix SAC and DDPG tensorflow policy can't do grad_clip (#10499)
* Fix sac_tf_policy clip_by_norm missing argument

* Fix ddpg_tf_policy clip_by_norm missing argument

* Fix format
2020-09-11 12:04:44 -07:00
..
tests [RLlib] Tf2.x native. (#8752) 2020-07-11 22:06:35 +02:00
__init__.py [RLlib] SAC Torch (incl. Atari learning) (#7984) 2020-04-15 13:25:16 +02:00
README.md [RLlib] Attention Net integration into ModelV2 and learning RL example. (#8371) 2020-05-18 17:26:40 +02:00
sac.py [RLlib] Deprecate old classes, methods, functions, config keys (in prep for RLlib 1.0). (#10544) 2020-09-06 10:58:00 +02:00
sac_tf_model.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
sac_tf_policy.py [rllib] Fix SAC and DDPG tensorflow policy can't do grad_clip (#10499) 2020-09-11 12:04:44 -07:00
sac_torch_model.py [Testing] Fix LINT/sphinx errors. (#8874) 2020-06-10 15:41:59 +02:00
sac_torch_policy.py [RLlib] SAC n_step > 1. (#10567) 2020-09-05 22:26:42 +02:00

Implementation of the Soft Actor-Critic algorithm:

[1] Soft Actor-Critic Algorithms and Applications - T. Haarnoja, A. Zhou, K. Hartikainen, et al. https://arxiv.org/abs/1812.05905.pdf

For supporting discrete action spaces, we implemented this patch on top of the original algorithm: [2] Soft Actor-Critic for Discrete Action Settings - Petros Christodoulou https://arxiv.org/pdf/1910.07207v2.pdf