.. |
env
|
[RLlib] Add coin game env. Matrix social dilemma env. With tests and examples. (#14208)
|
2021-03-09 17:26:20 +01:00 |
export
|
[Core] First pass at privatizing non-public Python APIs. (#14607)
|
2021-03-10 22:47:28 -08:00 |
models
|
[RLlib] Redo issue 14533 tf enable eager exec (#14984)
|
2021-03-29 20:07:44 +02:00 |
policy
|
[RLlib] Trajectory view API docs. (#12718)
|
2020-12-30 17:32:21 -08:00 |
serving
|
[RLlib] Extend on_learn_on_batch callback to allow for custom metrics to be added. (#13584)
|
2021-02-08 15:02:19 +01:00 |
simulators/sumo
|
[RLlib] Integration with SUMO Simulator (#11710)
|
2020-11-03 09:45:03 +01:00 |
__init__.py
|
[rllib] Try moving RLlib to top level dir (#5324)
|
2019-08-05 23:25:49 -07:00 |
attention_net.py
|
[RLlib] Fix 2 flakey test cases. (#14892)
|
2021-03-29 17:20:29 +02:00 |
attention_net_supervised.py
|
[RLlib] Support easy use_attention=True flag for using the GTrXL model. (#11698)
|
2021-01-01 14:06:23 -05:00 |
autoregressive_action_dist.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
batch_norm_model.py
|
[RLlib] Redo: Make TFModelV2 fully modular like TorchModelV2 (soft-deprecate register_variables, unify var names wrt torch). (#13363)
|
2021-01-14 14:44:33 +01:00 |
cartpole_lstm.py
|
[RLlib] Deprecate vf_share_layers in top-level PPO/MAML/MB-MPO configs. (#13397)
|
2021-01-19 09:51:35 +01:00 |
centralized_critic.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
centralized_critic_2.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
checkpoint_by_custom_criteria.py
|
[RLlib] Multi-GPU for tf-DQN/PG/A2C. (#13393)
|
2021-03-08 15:41:27 +01:00 |
coin_game_env.py
|
[RLlib] Add coin game env. Matrix social dilemma env. With tests and examples. (#14208)
|
2021-03-09 17:26:20 +01:00 |
complex_struct_space.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
custom_env.py
|
[RLlib] Add HowTo set env seed to our custom env example script. (#14471)
|
2021-03-16 08:12:27 +01:00 |
custom_eval.py
|
[RLlib] Fix inconsistency wrt batch size in SampleCollector (traj. view API). Makes DD-PPO work with traj. view API. (#12063)
|
2020-11-19 19:01:14 +01:00 |
custom_experiment.py
|
[RLlib] Example and test for custom Trainer wrapper experiments (#14652)
|
2021-03-24 16:22:46 +01:00 |
custom_fast_model.py
|
[RLlib] Fix flakey custom_fast_model_torch/tf tests. (#15330)
|
2021-04-15 16:10:29 +02:00 |
custom_keras_model.py
|
[RLlib] Redo issue 14533 tf enable eager exec (#14984)
|
2021-03-29 20:07:44 +02:00 |
custom_loss.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
custom_metrics_and_callbacks.py
|
[RLlib] Discussion 1513: on_episode_step() callback called after very first reset (should not). (#15218)
|
2021-04-11 13:16:17 +02:00 |
custom_metrics_and_callbacks_legacy.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
custom_model_api.py
|
[RLlib] Add more detailed Documentation on Model building API (#13261)
|
2021-01-09 12:38:29 +01:00 |
custom_observation_filters.py
|
[rllib] Rrk/12079 custom filters (#12095)
|
2020-11-19 13:20:20 -08:00 |
custom_rnn_model.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
custom_tf_policy.py
|
[RLlib] MB-MPO cleanup (comments, docstrings, type annotations). (#11033)
|
2020-10-06 20:28:16 +02:00 |
custom_torch_policy.py
|
[RLlib] JAXPolicy prep. PR #1. (#13077)
|
2020-12-26 20:14:18 -05:00 |
custom_train_fn.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
dmlab_watermaze.py
|
Remove future imports (#6724)
|
2020-01-09 00:15:48 -08:00 |
eager_execution.py
|
[RLlib] Redo issue 14533 tf enable eager exec (#14984)
|
2021-03-29 20:07:44 +02:00 |
env_rendering_and_recording.py
|
[RLlib] Fix env rendering and recording options (for non-local mode; >0 workers; +evaluation-workers). (#14796)
|
2021-03-23 10:06:06 +01:00 |
hierarchical_training.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
iterated_prisoners_dilemma_env.py
|
[RLlib] Add coin game env. Matrix social dilemma env. With tests and examples. (#14208)
|
2021-03-09 17:26:20 +01:00 |
lstm_auto_wrapping.py
|
[RLlib] Preparatory PR for: Documentation on Model Building. (#13260)
|
2021-01-08 10:56:09 +01:00 |
mobilenet_v2_with_lstm.py
|
[RLlib] Deprecate vf_share_layers in top-level PPO/MAML/MB-MPO configs. (#13397)
|
2021-01-19 09:51:35 +01:00 |
multi_agent_cartpole.py
|
[RLlib] Issue with agent-id -> pol-id mapping not required to be fixed across different episodes. (#15020)
|
2021-03-30 19:25:52 +02:00 |
multi_agent_custom_policy.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
multi_agent_independent_learning.py
|
[RLlib] Env directory cleanup and tests. (#13082)
|
2021-01-19 10:09:39 +01:00 |
multi_agent_parameter_sharing.py
|
[RLlib] Env directory cleanup and tests. (#13082)
|
2021-01-19 10:09:39 +01:00 |
multi_agent_two_trainers.py
|
[RLlib] Serve + RLlib example script. (#14416)
|
2021-03-03 14:33:03 +01:00 |
nested_action_spaces.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
parallel_evaluation_and_training.py
|
[RLlib] Support parallelizing evaluation and training (optional). (#15040)
|
2021-04-13 09:53:35 +02:00 |
parametric_actions_cartpole.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
pettingzoo_env.py
|
[RLlib] Allow rllib rollout to run distributed via evaluation workers. (#13718)
|
2021-02-08 12:05:16 +01:00 |
random_parametric_agent.py
|
[RLLib] Random Parametric Trainer (#11366)
|
2020-11-04 11:12:51 +01:00 |
rock_paper_scissors_multiagent.py
|
[client][rllib] Add client_mode_hook for ray.get_gpu_ids (#15185)
|
2021-04-08 23:36:11 -07:00 |
rollout_worker_custom_workflow.py
|
[RLlib] BC/MARWIL/recurrent nets minor cleanups and bug fixes. (#13064)
|
2020-12-27 09:46:03 -05:00 |
saving_experiences.py
|
[Core] First pass at privatizing non-public Python APIs. (#14607)
|
2021-03-10 22:47:28 -08:00 |
serve_and_rllib.py
|
[RLlib] Serve + RLlib example script. (#14416)
|
2021-03-03 14:33:03 +01:00 |
slateq.py
|
[RLlib] Implement the SlateQ algorithm (#11450)
|
2020-11-03 09:52:04 +01:00 |
sumo_env_local.py
|
[RLlib] Multi-GPU for tf-DQN/PG/A2C. (#13393)
|
2021-03-08 15:41:27 +01:00 |
trajectory_view_api.py
|
[RLlib] Trajectory view blogpost example script changes. (#14893)
|
2021-03-25 14:16:50 +01:00 |
two_step_game.py
|
[RLlib] Fix all example scripts to run on GPUs. (#11105)
|
2020-10-02 23:07:44 +02:00 |
two_trainer_workflow.py
|
[RLlib] Remove all non-trajectory view API code. (#14860)
|
2021-03-23 09:50:18 -07:00 |
unity3d_env_local.py
|
[RLlib] Unity3D example broken due to change in ML-Agents API. Attention-net prev-n-a/r. Attention-wrapper works with images. (#14569)
|
2021-03-12 18:27:25 +01:00 |
vizdoom_with_attention_net.py
|
[RLlib] Unity3D example broken due to change in ML-Agents API. Attention-net prev-n-a/r. Attention-wrapper works with images. (#14569)
|
2021-03-12 18:27:25 +01:00 |