ray/rllib/examples
Kai Fricke e1a7efe148
[tune] Use Checkpoint.to_bytes() for store_to_object (#25805)
We currently use our own serialization to ship checkpoints as objects. Instead we should use the Checkpoint class. This PR also adds support to create results from checkpoints pointing to object references.

Depends on #26351

Signed-off-by: Kai Fricke <kai@anyscale.com>
2022-07-08 18:01:20 +01:00
..
bandit [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
documentation [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
env Clean up docstyle in python modules and add LINT rule (#25272) 2022-06-01 11:27:54 -07:00
export [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
inference_and_serving [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
models Clean up docstyle in python modules and add LINT rule (#25272) 2022-06-01 11:27:54 -07:00
multi_agent_and_self_play Revert "Revert "[RLlib] AlphaStar: Parallelized, multi-agent/multi-GPU learni…" (#22153) 2022-02-08 16:43:00 +01:00
policy [RLlib] Memory leak finding toolset using tracemalloc + CI memory leak tests. (#15412) 2022-04-12 07:50:09 +02:00
serving [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
simulators/sumo Clean up docstyle in python modules and add LINT rule (#25272) 2022-06-01 11:27:54 -07:00
tune [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
action_masking.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
attention_net.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
attention_net_supervised.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
autoregressive_action_dist.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
bare_metal_policy_with_custom_view_reqs.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
batch_norm_model.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
cartpole_lstm.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00
centralized_critic.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
centralized_critic_2.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
checkpoint_by_custom_criteria.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
coin_game_env.py [RLlib] Move all remaining algos into algorithms directory. (#25366) 2022-06-04 07:35:24 +02:00
complex_struct_space.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
compute_adapted_gae_on_postprocess_trajectory.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
curriculum_learning.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_env.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_eval.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_experiment.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_fast_model.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
custom_input_api.py Clean up docstyle in python modules and add LINT rule (#25272) 2022-06-01 11:27:54 -07:00
custom_keras_model.py [RLlib] Agents to algos: DQN w/o Apex and R2D2, DDPG/TD3, SAC, SlateQ, QMIX, PG, Bandits (#24896) 2022-05-19 18:30:42 +02:00
custom_logger.py [tune] Next deprecation cycle (#24076) 2022-04-26 09:30:15 +01:00
custom_loss.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
custom_metrics_and_callbacks.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_metrics_and_callbacks_legacy.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
custom_model_api.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
custom_model_loss_and_metrics.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
custom_observation_filters.py [RLlib] Filter.clear_buffer() deprecated (use Filter.reset_buffer() instead). (#22246) 2022-02-10 02:58:43 +01:00
custom_rnn_model.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_tf_policy.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_torch_policy.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
custom_train_fn.py [RLlib] Move all remaining algos into algorithms directory. (#25366) 2022-06-04 07:35:24 +02:00
custom_vector_env.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
deterministic_training.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
dmlab_watermaze.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
eager_execution.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00
env_rendering_and_recording.py Clean up docstyle in python modules and add LINT rule (#25272) 2022-06-01 11:27:54 -07:00
fractional_gpus.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
hierarchical_training.py [tune] Next deprecation cycle (#24076) 2022-04-26 09:30:15 +01:00
iterated_prisoners_dilemma_env.py [RLlib] Move all remaining algos into algorithms directory. (#25366) 2022-06-04 07:35:24 +02:00
lstm_auto_wrapping.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
mobilenet_v2_with_lstm.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
multi_agent_cartpole.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
multi_agent_custom_policy.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
multi_agent_different_spaces_for_agents.py [RLlib] Discussion 6060 and 5120: auto-infer different agents' spaces in multi-agent env. (#24649) 2022-05-27 14:56:24 +02:00
multi_agent_independent_learning.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
multi_agent_parameter_sharing.py [RLlib] Trainer.training_iteration -> Trainer.training_step; Iterations vs reportings: Clarification of terms. (#25076) 2022-06-10 17:09:18 +02:00
multi_agent_two_trainers.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
nested_action_spaces.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
offline_rl.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00
parallel_evaluation_and_training.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
parametric_actions_cartpole.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
parametric_actions_cartpole_embeddings_learnt_by_model.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
partial_gpus.py [Lint] Cleanup incorrectly formatted strings (Part 1: RLLib). (#23128) 2022-03-15 17:34:21 +01:00
preprocessing_disabled.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
random_parametric_agent.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00
re3_exploration.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
recommender_system_with_recsim_and_slateq.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
remote_base_env_with_custom_api.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
remote_envs_with_inference_done_on_main_node.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00
replay_buffer_api.py [RLlib] Take replay buffer api example out of GPU examples. (#25841) 2022-06-16 19:12:38 +02:00
restore_1_of_n_agents_from_checkpoint.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
rnnsac_stateless_cartpole.py [tune] Use Checkpoint.to_bytes() for store_to_object (#25805) 2022-07-08 18:01:20 +01:00
rock_paper_scissors_multiagent.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
rollout_worker_custom_workflow.py [tune/structure] Introduce execution package (#26015) 2022-06-23 11:13:19 +01:00
saving_experiences.py [CI] Format Python code with Black (#21975) 2022-01-29 18:41:57 -08:00
sb2rllib_rllib_example.py [RLlib] Move all remaining algos into algorithms directory. (#25366) 2022-06-04 07:35:24 +02:00
sb2rllib_sb_example.py [RLlib] Examples for training, saving, loading, testing an agent with SB & RLlib (#15897) 2021-05-19 16:36:59 +02:00
self_play_league_based_with_open_spiel.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
self_play_with_open_spiel.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
sumo_env_local.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
trajectory_view_api.py [RLlib] Trainer to Algorithm renaming. (#25539) 2022-06-11 15:10:39 +02:00
two_step_game.py [RLlib] Agents to algos: DQN w/o Apex and R2D2, DDPG/TD3, SAC, SlateQ, QMIX, PG, Bandits (#24896) 2022-05-19 18:30:42 +02:00
two_trainer_workflow.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00
unity3d_env_local.py [RLlib] Issue 21489: Unity3D env lacks group rewards (#24016). 2022-04-21 18:49:52 +02:00
vizdoom_with_attention_net.py [RLlib] More Trainer -> Algorithm renaming cleanups. (#25869) 2022-06-20 15:54:00 +02:00