mirror of
https://github.com/vale981/ray
synced 2025-03-06 18:41:40 -05:00
24 lines
866 B
Python
24 lines
866 B
Python
import ray.rllib.agents.ppo.appo as appo # noqa
|
|
from ray.rllib.algorithms.ppo.ppo import PPOConfig, PPO as PPOTrainer, DEFAULT_CONFIG
|
|
from ray.rllib.algorithms.ppo.ppo_tf_policy import PPOTF1Policy, PPOTF2Policy
|
|
from ray.rllib.algorithms.ppo.ppo_torch_policy import PPOTorchPolicy
|
|
from ray.rllib.algorithms.appo.appo import APPOConfig, APPO as APPOTrainer
|
|
from ray.rllib.algorithms.appo.appo_tf_policy import APPOTF1Policy, APPOTF2Policy
|
|
from ray.rllib.algorithms.appo.appo_torch_policy import APPOTorchPolicy
|
|
from ray.rllib.algorithms.ddppo.ddppo import DDPPOConfig, DDPPO as DDPPOTrainer
|
|
|
|
__all__ = [
|
|
"APPOConfig",
|
|
"APPOTF1Policy",
|
|
"APPOTF2Policy",
|
|
"APPOTorchPolicy",
|
|
"APPOTrainer",
|
|
"DDPPOConfig",
|
|
"DDPPOTrainer",
|
|
"DEFAULT_CONFIG",
|
|
"PPOConfig",
|
|
"PPOTF1Policy",
|
|
"PPOTF2Policy",
|
|
"PPOTorchPolicy",
|
|
"PPOTrainer",
|
|
]
|