..
atari-ddppo.yaml
[RLlib] Deprecate vf_share_layers
in top-level PPO/MAML/MB-MPO configs. ( #13397 )
2021-01-19 09:51:35 +01:00
atari-ppo.yaml
[RLlib] Deprecate vf_share_layers
in top-level PPO/MAML/MB-MPO configs. ( #13397 )
2021-01-19 09:51:35 +01:00
cartpole-appo-vtrace-fake-gpus.yaml
[RLlib] Optionally don't drop last ts in v-trace calculations (APPO and IMPALA). ( #19601 )
2021-11-03 10:01:34 +01:00
cartpole-appo-vtrace-separate-losses.yaml
[RLlib] Add all simple learning tests as framework=tf2
. ( #19273 )
2021-11-02 12:10:17 +01:00
cartpole-appo-vtrace.yaml
[RLlib] Minor fixes/cleanups; chop_into_sequences now handles nested data. ( #19408 )
2021-11-05 14:39:28 +01:00
cartpole-appo.yaml
[RLlib] Refactor: All tf static graph code should reside inside Policy class. ( #17169 )
2021-07-20 14:58:13 -04:00
cartpole-ddppo.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
cartpole-grid-search-example.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
cartpole-ppo-fake-gpus.yaml
[RLlib] Move existing fake multi-GPU learning tests into separate buildkite job. ( #18065 )
2021-08-31 14:56:53 +02:00
cartpole-ppo-hyperband.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
cartpole-ppo.yaml
[RLlib] Deprecate vf_share_layers
in top-level PPO/MAML/MB-MPO configs. ( #13397 )
2021-01-19 09:51:35 +01:00
frozenlake-appo-vtrace.yaml
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. ( #19535 )
2021-11-03 16:24:00 +01:00
halfcheetah-appo.yaml
[RLlib] Refactor: All tf static graph code should reside inside Policy class. ( #17169 )
2021-07-20 14:58:13 -04:00
halfcheetah-ppo.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
hopper-ppo.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
humanoid-ppo-gae.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
humanoid-ppo.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00
memory-leak-test-appo.yaml
[RLlib] Memory leak finding toolset using tracemalloc + CI memory leak tests. ( #15412 )
2022-04-12 07:50:09 +02:00
memory-leak-test-ppo.yaml
[RLlib] Memory leak finding toolset using tracemalloc + CI memory leak tests. ( #15412 )
2022-04-12 07:50:09 +02:00
pendulum-appo.yaml
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. ( #19535 )
2021-11-03 16:24:00 +01:00
pendulum-ddppo.yaml
[RLlib] DD-PPO training iteration fn. ( #24118 )
2022-04-22 15:22:14 -07:00
pendulum-ppo.yaml
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. ( #19535 )
2021-11-03 16:24:00 +01:00
pendulum-transformed-actions-ppo.yaml
[RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes ( #16531 )
2021-06-30 12:32:11 +02:00
pong-appo.yaml
[RLlib] Refactor: All tf static graph code should reside inside Policy class. ( #17169 )
2021-07-20 14:58:13 -04:00
pong-ppo.yaml
[RLlib] Deprecate vf_share_layers
in top-level PPO/MAML/MB-MPO configs. ( #13397 )
2021-01-19 09:51:35 +01:00
recomm-sys001-ppo.yaml
[RLlib] Slate-Q tf implementation and tests/benchmarks. ( #22389 )
2022-02-22 09:36:44 +01:00
repeatafterme-ppo-lstm.yaml
[RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). ( #18591 )
2021-09-15 22:16:48 +02:00
unity3d-soccer-strikers-vs-goalie-ppo.yaml
[RLlib] Env directory cleanup and tests. ( #13082 )
2021-01-19 10:09:39 +01:00
walker2d-ppo.yaml
[RLlib] Auto-framework, retire use_pytorch
in favor of framework=...
( #8520 )
2020-05-27 16:19:13 +02:00