2020-05-26 11:10:27 +02:00
|
|
|
# Can expect improvement to -140 reward in ~300-500k timesteps.
|
|
|
|
pendulum-ppo:
|
2018-07-02 08:20:53 -07:00
|
|
|
env: Pendulum-v0
|
|
|
|
run: PPO
|
|
|
|
stop:
|
2020-04-30 15:48:11 +02:00
|
|
|
episode_reward_mean: -500
|
|
|
|
timesteps_total: 400000
|
2018-07-02 08:20:53 -07:00
|
|
|
config:
|
2020-05-26 11:10:27 +02:00
|
|
|
# Works for both torch and tf.
|
2020-05-27 16:19:13 +02:00
|
|
|
framework: tf
|
2020-06-05 08:34:21 +02:00
|
|
|
train_batch_size: 512
|
2018-09-23 13:11:17 -07:00
|
|
|
vf_clip_param: 10.0
|
2018-12-11 17:21:53 -08:00
|
|
|
num_workers: 0
|
2020-06-05 08:34:21 +02:00
|
|
|
num_envs_per_worker: 20
|
2018-07-02 08:20:53 -07:00
|
|
|
lambda: 0.1
|
|
|
|
gamma: 0.95
|
2018-09-05 12:06:13 -07:00
|
|
|
lr: 0.0003
|
|
|
|
sgd_minibatch_size: 64
|
2020-06-05 08:34:21 +02:00
|
|
|
num_sgd_iter: 6
|
2019-03-01 13:19:33 -08:00
|
|
|
observation_filter: MeanStdFilter
|