ray/rllib/examples
2021-11-12 12:33:39 -08:00
..
documentation [RLlib; documentation] "RLlib in 60sec" overhaul. (#20215) 2021-11-10 22:20:06 +01:00
env [RLlib] Tf2 + eager-tracing same speed as framework=tf; Add more test coverage for tf2+tracing. (#19981) 2021-11-05 16:10:00 +01:00
export [RLlib] Fix failing test cases: Soft-deprecate ModelV2.from_batch (in favor of ModelV2.__call__). (#19693) 2021-10-25 15:00:00 +02:00
inference_and_serving [RLlib] Issue 20062: Action inference examples missing (#20144) 2021-11-10 18:49:06 +01:00
models [RLlib] Tf2 + eager-tracing same speed as framework=tf; Add more test coverage for tf2+tracing. (#19981) 2021-11-05 16:10:00 +01:00
policy [RLlib] No Preprocessors (part 2). (#18468) 2021-09-23 12:56:45 +02:00
serving [RLlib] Issue 18668: Unity3D env client/server example not working (fix + add to test cases). (#18942) 2021-09-30 08:30:20 +02:00
simulators/sumo [RLlib] Integration with SUMO Simulator (#11710) 2020-11-03 09:45:03 +01:00
tune [Release] Refactor User Tests (#20028) 2021-11-05 17:28:37 -07:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
action_masking.py [RLlib] Add simple action-masking example script/env/model (tf and torch). (#18494) 2021-09-11 23:08:09 +02:00
attention_net.py [RLlib] Better PolicyServer example (w/ or w/o tune) and add printing out actual listen port address in log-level=INFO. (#18254) 2021-08-31 22:03:23 +02:00
attention_net_supervised.py [RLlib] Support easy use_attention=True flag for using the GTrXL model. (#11698) 2021-01-01 14:06:23 -05:00
autoregressive_action_dist.py [RLlib] Better example scripts: Description --no-tune and --local-mode CLI options (autoregressive_action_dist.py) (#17705) 2021-08-16 22:08:13 +02:00
bare_metal_policy_with_custom_view_reqs.py [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
batch_norm_model.py [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
cartpole_lstm.py [RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531) 2021-06-30 12:32:11 +02:00
centralized_critic.py [RLlib] Fix deprecated warning for torch_ops.py (soft-replaced by torch_utils.py). (#19982) 2021-11-03 10:00:46 +01:00
centralized_critic_2.py [RLlib] Re-do: Trainer: Support add and delete Policies. (#16569) 2021-06-21 13:46:01 +02:00
checkpoint_by_custom_criteria.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
coin_game_env.py [RLlib] Re-do: Trainer: Support add and delete Policies. (#16569) 2021-06-21 13:46:01 +02:00
complex_struct_space.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
curriculum_learning.py [RLlib] Add simple curriculum learning API and example script. (#15740) 2021-05-16 17:35:10 +02:00
custom_env.py [RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591) 2021-09-15 22:16:48 +02:00
custom_eval.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
custom_experiment.py [RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531) 2021-06-30 12:32:11 +02:00
custom_fast_model.py [RLlib] Refactor: All tf static graph code should reside inside Policy class. (#17169) 2021-07-20 14:58:13 -04:00
custom_input_api.py [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
custom_keras_model.py [RLlib] Unify all RLlib Trainer.train() -> results[info][learner][policy ID][learner_stats] and add structure tests. (#18879) 2021-09-30 16:39:05 +02:00
custom_logger.py [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
custom_loss.py [RLlib] Fix ModelV2 custom metrics for torch. (#16734) 2021-07-01 13:01:40 +02:00
custom_metrics_and_callbacks.py [RLlib; Docs overhaul] Docstring cleanup: Evaluation (#19783) 2021-10-29 12:03:56 +02:00
custom_metrics_and_callbacks_legacy.py [RLlib] Fix all example scripts to run on GPUs. (#11105) 2020-10-02 23:07:44 +02:00
custom_model_api.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
custom_model_loss_and_metrics.py [RLlib] Unify all RLlib Trainer.train() -> results[info][learner][policy ID][learner_stats] and add structure tests. (#18879) 2021-09-30 16:39:05 +02:00
custom_observation_filters.py [RLlib] No Preprocessors; preparatory PR #1 (#18367) 2021-09-09 08:10:42 +02:00
custom_rnn_model.py [RLlib] CQL BC loss fixes; PPO/PG/A2|3C action normalization fixes (#16531) 2021-06-30 12:32:11 +02:00
custom_tf_policy.py [RLlib] Fix failing test cases: Soft-deprecate ModelV2.from_batch (in favor of ModelV2.__call__). (#19693) 2021-10-25 15:00:00 +02:00
custom_torch_policy.py [RLlib] JAXPolicy prep. PR #1. (#13077) 2020-12-26 20:14:18 -05:00
custom_train_fn.py [RLlib Testig] Split and unflake more CI tests (make sure all jobs are < 30min). (#18591) 2021-09-15 22:16:48 +02:00
custom_vector_env.py [RLlib] Discussion 2294: Custom vector env example and fix. (#16083) 2021-07-28 10:40:04 -04:00
deterministic_training.py [rllib] Add deterministic test to gpu (#19306) 2021-10-26 10:11:39 -07:00
dmlab_watermaze.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
eager_execution.py [RLlib] Fix failing test cases: Soft-deprecate ModelV2.from_batch (in favor of ModelV2.__call__). (#19693) 2021-10-25 15:00:00 +02:00
env_rendering_and_recording.py [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
fractional_gpus.py [RLlib] Fix ModelV2 custom metrics for torch. (#16734) 2021-07-01 13:01:40 +02:00
hierarchical_training.py [RLlib] Add worker arg (optional) to policy_mapping_fn. (#18184) 2021-09-17 12:07:11 +02:00
iterated_prisoners_dilemma_env.py [RLlib] Re-do: Trainer: Support add and delete Policies. (#16569) 2021-06-21 13:46:01 +02:00
lstm_auto_wrapping.py [RLlib] Preparatory PR for: Documentation on Model Building. (#13260) 2021-01-08 10:56:09 +01:00
mobilenet_v2_with_lstm.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
multi_agent_cartpole.py [RLlib] Add worker arg (optional) to policy_mapping_fn. (#18184) 2021-09-17 12:07:11 +02:00
multi_agent_custom_policy.py [RLlib] Redo simplify multi agent config dict: Reverted b/c seemed to break test_typing (non RLlib test). (#17046) 2021-07-15 05:51:24 -04:00
multi_agent_independent_learning.py [RLlib] Redo simplify multi agent config dict: Reverted b/c seemed to break test_typing (non RLlib test). (#17046) 2021-07-15 05:51:24 -04:00
multi_agent_parameter_sharing.py [RLlib] Redo simplify multi agent config dict: Reverted b/c seemed to break test_typing (non RLlib test). (#17046) 2021-07-15 05:51:24 -04:00
multi_agent_two_trainers.py [RLlib] Add worker arg (optional) to policy_mapping_fn. (#18184) 2021-09-17 12:07:11 +02:00
nested_action_spaces.py [RLlib] Fix crash when using StochasticSampling exploration (most PG-style algos) w/ tf and numpy > 1.19.5 (#18366) 2021-09-06 12:14:00 +02:00
offline_rl.py [RLlib] Upgrade gym version to 0.21 and deprecate pendulum-v0. (#19535) 2021-11-03 16:24:00 +01:00
parallel_evaluation_and_training.py [RLlib] Add support for evaluation_num_episodes=auto (run eval for as long as the parallel train step takes). (#18380) 2021-09-07 08:08:37 +02:00
parametric_actions_cartpole.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
parametric_actions_cartpole_embeddings_learnt_by_model.py [RLlib] New and changed version of parametric actions cartpole example + small suggested update in policy_client.py (#15664) 2021-07-28 15:25:09 -04:00
partial_gpus.py [RLlib] Fix ModelV2 custom metrics for torch. (#16734) 2021-07-01 13:01:40 +02:00
preprocessing_disabled.py [RLlib] No Preprocessors (part 2). (#18468) 2021-09-23 12:56:45 +02:00
random_parametric_agent.py [RLlib] Unify the way we create local replay buffer for all agents (#19627) 2021-10-26 20:56:02 +02:00
recsim_with_slateq.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
remote_envs_with_inference_done_on_main_node.py [RLlib] Redo #17410: Example script: Remote worker envs with inference done on main node. (#17960) 2021-08-20 08:02:18 +02:00
remote_vector_env_with_custom_api.py [RLlib] Issue 18104: Cannot set remote_worker_envs=True for non local-mode and MultiAgentEnv. (#19133) 2021-10-07 22:39:21 +02:00
restore_1_of_n_agents_from_checkpoint.py [RLlib] Add worker arg (optional) to policy_mapping_fn. (#18184) 2021-09-17 12:07:11 +02:00
rnnsac_stateless_cartpole.py [RLlib] Add RNN-SAC agent (#16577) 2021-07-25 10:04:52 -04:00
rock_paper_scissors_multiagent.py Revert "[RLlib] POC: PGTrainer class that works by sub-classing, not trainer_template.py. (#20055)" (#20284) 2021-11-12 13:09:43 +00:00
rollout_worker_custom_workflow.py [Tune] Remove legacy resources implementations in Runner and Executor. (#19773) 2021-11-12 12:33:39 -08:00
saving_experiences.py [Core] First pass at privatizing non-public Python APIs. (#14607) 2021-03-10 22:47:28 -08:00
sb2rllib_rllib_example.py [RLlib] Better example scripts: Description --no-tune and --local-mode CLI options (#17038) 2021-07-26 22:25:48 -04:00
sb2rllib_sb_example.py [RLlib] Examples for training, saving, loading, testing an agent with SB & RLlib (#15897) 2021-05-19 16:36:59 +02:00
self_play_league_based_with_open_spiel.py [RLlib] Add worker arg (optional) to policy_mapping_fn. (#18184) 2021-09-17 12:07:11 +02:00
self_play_with_open_spiel.py [RLlib] Add worker arg (optional) to policy_mapping_fn. (#18184) 2021-09-17 12:07:11 +02:00
sumo_env_local.py [RLlib] Re-do: Trainer: Support add and delete Policies. (#16569) 2021-06-21 13:46:01 +02:00
trajectory_view_api.py [RLlib] Custom view requirements (e.g. for prev-n-obs) work with compute_single_action and compute_actions_from_input_dict. (#18921) 2021-09-30 15:03:37 +02:00
two_step_game.py [RLlib] POC: Separate losses for APPO/IMPALA. Enable TFPolicy to handle multiple optimizers/losses (like TorchPolicy). (#18669) 2021-09-21 22:00:14 +02:00
two_trainer_workflow.py [RLlib] Some minor cleanups (buffer buffer_size -> capacity and others). (#19623) 2021-10-25 09:42:39 +02:00
unity3d_env_local.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00
vizdoom_with_attention_net.py [RLlib] Examples scripts add argparse help and replace --torch with --framework. (#15832) 2021-05-18 13:18:12 +02:00