2022-06-11 15:10:39 +02:00
|
|
|
import ray.rllib.agents.ppo.appo as appo # noqa
|
2022-06-04 07:35:24 +02:00
|
|
|
from ray.rllib.algorithms.ppo.ppo import PPOConfig, PPO as PPOTrainer, DEFAULT_CONFIG
|
|
|
|
from ray.rllib.algorithms.ppo.ppo_tf_policy import PPOTF1Policy, PPOTF2Policy
|
|
|
|
from ray.rllib.algorithms.ppo.ppo_torch_policy import PPOTorchPolicy
|
|
|
|
from ray.rllib.algorithms.appo.appo import APPOConfig, APPO as APPOTrainer
|
|
|
|
from ray.rllib.algorithms.appo.appo_tf_policy import APPOTF1Policy, APPOTF2Policy
|
|
|
|
from ray.rllib.algorithms.appo.appo_torch_policy import APPOTorchPolicy
|
|
|
|
from ray.rllib.algorithms.ddppo.ddppo import DDPPOConfig, DDPPO as DDPPOTrainer
|
2018-07-01 00:05:08 -07:00
|
|
|
|
2020-05-08 08:20:18 +02:00
|
|
|
__all__ = [
|
2022-05-02 15:06:23 +02:00
|
|
|
"APPOConfig",
|
2022-06-04 07:35:24 +02:00
|
|
|
"APPOTF1Policy",
|
|
|
|
"APPOTF2Policy",
|
|
|
|
"APPOTorchPolicy",
|
2020-05-08 08:20:18 +02:00
|
|
|
"APPOTrainer",
|
2022-05-22 13:05:24 +02:00
|
|
|
"DDPPOConfig",
|
2020-05-08 08:20:18 +02:00
|
|
|
"DDPPOTrainer",
|
|
|
|
"DEFAULT_CONFIG",
|
2022-03-31 18:26:12 +02:00
|
|
|
"PPOConfig",
|
2022-06-04 07:35:24 +02:00
|
|
|
"PPOTF1Policy",
|
|
|
|
"PPOTF2Policy",
|
2020-05-08 08:20:18 +02:00
|
|
|
"PPOTorchPolicy",
|
|
|
|
"PPOTrainer",
|
|
|
|
]
|