ray/rllib/agents/sac
2020-07-11 22:06:35 +02:00
..
tests [RLlib] Tf2.x native. (#8752) 2020-07-11 22:06:35 +02:00
__init__.py [RLlib] SAC Torch (incl. Atari learning) (#7984) 2020-04-15 13:25:16 +02:00
README.md [RLlib] Attention Net integration into ModelV2 and learning RL example. (#8371) 2020-05-18 17:26:40 +02:00
sac.py [RLlib] DDPG and SAC eager support (preparation for tf2.x) (#9204) 2020-07-08 16:12:20 +02:00
sac_tf_model.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
sac_tf_policy.py [RLlib] Tf2.x native. (#8752) 2020-07-11 22:06:35 +02:00
sac_torch_model.py [Testing] Fix LINT/sphinx errors. (#8874) 2020-06-10 15:41:59 +02:00
sac_torch_policy.py [RLlib] Minor cleanup in preparation to tf2.x support. (#9130) 2020-06-25 19:01:32 +02:00

Implementation of the Soft Actor-Critic algorithm:

[1] Soft Actor-Critic Algorithms and Applications - T. Haarnoja, A. Zhou, K. Hartikainen, et al. https://arxiv.org/abs/1812.05905.pdf

For supporting discrete action spaces, we implemented this patch on top of the original algorithm: [2] Soft Actor-Critic for Discrete Action Settings - Petros Christodoulou https://arxiv.org/pdf/1910.07207v2.pdf