ray/rllib/agents/ppo/__init__.py

24 lines
866 B
Python

import ray.rllib.agents.ppo.appo as appo # noqa
from ray.rllib.algorithms.ppo.ppo import PPOConfig, PPO as PPOTrainer, DEFAULT_CONFIG
from ray.rllib.algorithms.ppo.ppo_tf_policy import PPOTF1Policy, PPOTF2Policy
from ray.rllib.algorithms.ppo.ppo_torch_policy import PPOTorchPolicy
from ray.rllib.algorithms.appo.appo import APPOConfig, APPO as APPOTrainer
from ray.rllib.algorithms.appo.appo_tf_policy import APPOTF1Policy, APPOTF2Policy
from ray.rllib.algorithms.appo.appo_torch_policy import APPOTorchPolicy
from ray.rllib.algorithms.ddppo.ddppo import DDPPOConfig, DDPPO as DDPPOTrainer
__all__ = [
"APPOConfig",
"APPOTF1Policy",
"APPOTF2Policy",
"APPOTorchPolicy",
"APPOTrainer",
"DDPPOConfig",
"DDPPOTrainer",
"DEFAULT_CONFIG",
"PPOConfig",
"PPOTF1Policy",
"PPOTF2Policy",
"PPOTorchPolicy",
"PPOTrainer",
]