.. |
atari-ddppo.yaml
|
[RLlib] Deprecate vf_share_layers in top-level PPO/MAML/MB-MPO configs. (#13397)
|
2021-01-19 09:51:35 +01:00 |
atari-ppo.yaml
|
[RLlib] Deprecate vf_share_layers in top-level PPO/MAML/MB-MPO configs. (#13397)
|
2021-01-19 09:51:35 +01:00 |
cartpole-appo-vtrace-fake-gpus.yaml
|
[RLlib] Optionally don't drop last ts in v-trace calculations (APPO and IMPALA). (#19601)
|
2021-11-03 10:01:34 +01:00 |
cartpole-appo-vtrace-separate-losses.yaml
|
[RLlib] Add all simple learning tests as framework=tf2 . (#19273)
|
2021-11-02 12:10:17 +01:00 |
cartpole-appo-vtrace.yaml
|
[RLlib] Minor fixes/cleanups; chop_into_sequences now handles nested data. (#19408)
|
2021-11-05 14:39:28 +01:00 |
cartpole-appo.yaml
|
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
2021-07-20 14:58:13 -04:00 |
cartpole-ddppo.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
cartpole-grid-search-example.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
cartpole-ppo-fake-gpus.yaml
|
[RLlib] Move existing fake multi-GPU learning tests into separate buildkite job. (#18065)
|
2021-08-31 14:56:53 +02:00 |
cartpole-ppo-hyperband.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
cartpole-ppo.yaml
|
[RLlib] Deprecate vf_share_layers in top-level PPO/MAML/MB-MPO configs. (#13397)
|
2021-01-19 09:51:35 +01:00 |
frozenlake-appo-vtrace.yaml
|
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
2021-11-03 16:24:00 +01:00 |
halfcheetah-appo.yaml
|
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
2021-07-20 14:58:13 -04:00 |
halfcheetah-ppo.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
hopper-ppo.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
humanoid-ppo-gae.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
humanoid-ppo.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |
memory-leak-test-appo.yaml
|
[RLlib] Memory leak finding toolset using tracemalloc + CI memory leak tests. (#15412)
|
2022-04-12 07:50:09 +02:00 |
memory-leak-test-ppo.yaml
|
[RLlib] Memory leak finding toolset using tracemalloc + CI memory leak tests. (#15412)
|
2022-04-12 07:50:09 +02:00 |
pendulum-appo.yaml
|
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
2021-11-03 16:24:00 +01:00 |
pendulum-ddppo.yaml
|
Revert revert #23906 [RLlib] DD-PPO training iteration function implementation. (#24035)
|
2022-04-21 17:37:49 +02:00 |
pendulum-ppo.yaml
|
[RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535)
|
2021-11-03 16:24:00 +01:00 |
pendulum-transformed-actions-ppo.yaml
|
[RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531)
|
2021-06-30 12:32:11 +02:00 |
pong-appo.yaml
|
[RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169)
|
2021-07-20 14:58:13 -04:00 |
pong-ppo.yaml
|
[RLlib] Deprecate vf_share_layers in top-level PPO/MAML/MB-MPO configs. (#13397)
|
2021-01-19 09:51:35 +01:00 |
recomm-sys001-ppo.yaml
|
[RLlib] Slate-Q tf implementation and tests/benchmarks. (#22389)
|
2022-02-22 09:36:44 +01:00 |
repeatafterme-ppo-lstm.yaml
|
[RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591)
|
2021-09-15 22:16:48 +02:00 |
unity3d-soccer-strikers-vs-goalie-ppo.yaml
|
[RLlib] Env directory cleanup and tests. (#13082)
|
2021-01-19 10:09:39 +01:00 |
walker2d-ppo.yaml
|
[RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520)
|
2020-05-27 16:19:13 +02:00 |