2018-06-09 16:46:23 -07:00
|
|
|
# This configuration can expect to reach 2000 reward in 150k-200k timesteps
|
|
|
|
halfcheetah-ddpg:
|
|
|
|
env: HalfCheetah-v2
|
|
|
|
run: DDPG
|
|
|
|
stop:
|
|
|
|
episode_reward_mean: 2000
|
|
|
|
time_total_s: 5400 # 90 minutes
|
|
|
|
config:
|
2020-05-27 16:19:13 +02:00
|
|
|
# Works for both torch and tf.
|
|
|
|
framework: tf
|
2018-06-09 16:46:23 -07:00
|
|
|
# === Model ===
|
|
|
|
actor_hiddens: [64, 64]
|
|
|
|
critic_hiddens: [64, 64]
|
|
|
|
n_step: 1
|
|
|
|
model: {}
|
|
|
|
gamma: 0.99
|
|
|
|
env_config: {}
|
|
|
|
|
|
|
|
# === Exploration ===
|
2020-03-01 20:53:35 +01:00
|
|
|
exploration_config:
|
|
|
|
initial_scale: 1.0
|
|
|
|
final_scale: 0.02
|
|
|
|
scale_timesteps: 10000
|
|
|
|
ou_base_scale: 0.1
|
|
|
|
ou_theta: 0.15
|
|
|
|
ou_sigma: 0.2
|
|
|
|
|
2018-06-09 16:46:23 -07:00
|
|
|
timesteps_per_iteration: 1000
|
|
|
|
target_network_update_freq: 0
|
|
|
|
tau: 0.001
|
|
|
|
|
|
|
|
# === Replay buffer ===
|
|
|
|
buffer_size: 10000
|
|
|
|
prioritized_replay: True
|
|
|
|
prioritized_replay_alpha: 0.6
|
|
|
|
prioritized_replay_beta: 0.4
|
|
|
|
prioritized_replay_eps: 0.000001
|
|
|
|
clip_rewards: False
|
|
|
|
|
|
|
|
# === Optimization ===
|
2019-04-26 17:49:53 -07:00
|
|
|
actor_lr: 0.001
|
|
|
|
critic_lr: 0.001
|
2021-08-26 14:09:20 +02:00
|
|
|
use_huber: false
|
2018-06-09 16:46:23 -07:00
|
|
|
huber_threshold: 1.0
|
|
|
|
l2_reg: 0.000001
|
|
|
|
learning_starts: 500
|
2020-03-14 12:05:04 -07:00
|
|
|
rollout_fragment_length: 1
|
2018-06-09 16:46:23 -07:00
|
|
|
train_batch_size: 64
|
|
|
|
|
|
|
|
# === Parallelism ===
|
|
|
|
num_workers: 0
|
|
|
|
num_gpus_per_worker: 0
|
2021-08-26 14:09:20 +02:00
|
|
|
worker_side_prioritization: false
|
2019-04-26 17:49:53 -07:00
|
|
|
|
|
|
|
# === Evaluation ===
|
|
|
|
evaluation_interval: 5
|
|
|
|
evaluation_num_episodes: 10
|