ray/rllib/tuned_examples/pendulum-ppo.yaml

19 lines
483 B
YAML
Raw Normal View History

# can expect improvement to -140 reward in ~300-500k timesteps
pendulum-ppo:
env: Pendulum-v0
run: PPO
config:
train_batch_size: 2048
vf_clip_param: 10.0
num_workers: 0
num_envs_per_worker: 10
lambda: 0.1
gamma: 0.95
lr: 0.0003
sgd_minibatch_size: 64
num_sgd_iter: 10
model:
fcnet_hiddens: [64, 64]
batch_mode: complete_episodes
observation_filter: MeanStdFilter