..
bandit
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
documentation
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
env
Clean up docstyle in python modules and add LINT rule ( #25272 )
2022-06-01 11:27:54 -07:00
export
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
inference_and_serving
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
models
Clean up docstyle in python modules and add LINT rule ( #25272 )
2022-06-01 11:27:54 -07:00
multi_agent_and_self_play
Revert "Revert "[RLlib] AlphaStar: Parallelized, multi-agent/multi-GPU learni…" ( #22153 )
2022-02-08 16:43:00 +01:00
policy
[RLlib] Memory leak finding toolset using tracemalloc + CI memory leak tests. ( #15412 )
2022-04-12 07:50:09 +02:00
serving
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
simulators /sumo
Clean up docstyle in python modules and add LINT rule ( #25272 )
2022-06-01 11:27:54 -07:00
tune
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
__init__.py
[rllib] Try moving RLlib to top level dir ( #5324 )
2019-08-05 23:25:49 -07:00
action_masking.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
attention_net.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
attention_net_supervised.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
autoregressive_action_dist.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
bare_metal_policy_with_custom_view_reqs.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
batch_norm_model.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
cartpole_lstm.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00
centralized_critic.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
centralized_critic_2.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
checkpoint_by_custom_criteria.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
coin_game_env.py
[RLlib] Move all remaining algos into algorithms
directory. ( #25366 )
2022-06-04 07:35:24 +02:00
complex_struct_space.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
compute_adapted_gae_on_postprocess_trajectory.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
curriculum_learning.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_env.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_eval.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_experiment.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_fast_model.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
custom_input_api.py
Clean up docstyle in python modules and add LINT rule ( #25272 )
2022-06-01 11:27:54 -07:00
custom_keras_model.py
[RLlib] Agents to algos: DQN w/o Apex and R2D2, DDPG/TD3, SAC, SlateQ, QMIX, PG, Bandits ( #24896 )
2022-05-19 18:30:42 +02:00
custom_logger.py
[tune] Next deprecation cycle ( #24076 )
2022-04-26 09:30:15 +01:00
custom_loss.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
custom_metrics_and_callbacks.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_metrics_and_callbacks_legacy.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
custom_model_api.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
custom_model_loss_and_metrics.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
custom_observation_filters.py
[RLlib] Filter.clear_buffer() deprecated (use Filter.reset_buffer() instead). ( #22246 )
2022-02-10 02:58:43 +01:00
custom_rnn_model.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_tf_policy.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_torch_policy.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
custom_train_fn.py
[RLlib] Move all remaining algos into algorithms
directory. ( #25366 )
2022-06-04 07:35:24 +02:00
custom_vector_env.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
deterministic_training.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
dmlab_watermaze.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
eager_execution.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00
env_rendering_and_recording.py
Clean up docstyle in python modules and add LINT rule ( #25272 )
2022-06-01 11:27:54 -07:00
fractional_gpus.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
hierarchical_training.py
[tune] Next deprecation cycle ( #24076 )
2022-04-26 09:30:15 +01:00
iterated_prisoners_dilemma_env.py
[RLlib] Move all remaining algos into algorithms
directory. ( #25366 )
2022-06-04 07:35:24 +02:00
lstm_auto_wrapping.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
mobilenet_v2_with_lstm.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
multi_agent_cartpole.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
multi_agent_custom_policy.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
multi_agent_different_spaces_for_agents.py
[RLlib] Discussion 6060 and 5120: auto-infer different agents' spaces in multi-agent env. ( #24649 )
2022-05-27 14:56:24 +02:00
multi_agent_independent_learning.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
multi_agent_parameter_sharing.py
[RLlib] Trainer.training_iteration -> Trainer.training_step; Iterations vs reportings: Clarification of terms. ( #25076 )
2022-06-10 17:09:18 +02:00
multi_agent_two_trainers.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
nested_action_spaces.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
offline_rl.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00
parallel_evaluation_and_training.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
parametric_actions_cartpole.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
parametric_actions_cartpole_embeddings_learnt_by_model.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
partial_gpus.py
[Lint] Cleanup incorrectly formatted strings (Part 1: RLLib). ( #23128 )
2022-03-15 17:34:21 +01:00
preprocessing_disabled.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
random_parametric_agent.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00
re3_exploration.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
recommender_system_with_recsim_and_slateq.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
remote_base_env_with_custom_api.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
remote_envs_with_inference_done_on_main_node.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00
replay_buffer_api.py
[RLlib] Take replay buffer api example out of GPU examples. ( #25841 )
2022-06-16 19:12:38 +02:00
restore_1_of_n_agents_from_checkpoint.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
rnnsac_stateless_cartpole.py
[tune] Use Checkpoint.to_bytes()
for store_to_object ( #25805 )
2022-07-08 18:01:20 +01:00
rock_paper_scissors_multiagent.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
rollout_worker_custom_workflow.py
[tune/structure] Introduce execution package ( #26015 )
2022-06-23 11:13:19 +01:00
saving_experiences.py
[CI] Format Python code with Black ( #21975 )
2022-01-29 18:41:57 -08:00
sb2rllib_rllib_example.py
[RLlib] Move all remaining algos into algorithms
directory. ( #25366 )
2022-06-04 07:35:24 +02:00
sb2rllib_sb_example.py
[RLlib] Examples for training, saving, loading, testing an agent with SB & RLlib ( #15897 )
2021-05-19 16:36:59 +02:00
self_play_league_based_with_open_spiel.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
self_play_with_open_spiel.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
sumo_env_local.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
trajectory_view_api.py
[RLlib] Trainer
to Algorithm
renaming. ( #25539 )
2022-06-11 15:10:39 +02:00
two_step_game.py
[RLlib] Agents to algos: DQN w/o Apex and R2D2, DDPG/TD3, SAC, SlateQ, QMIX, PG, Bandits ( #24896 )
2022-05-19 18:30:42 +02:00
two_trainer_workflow.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00
unity3d_env_local.py
[RLlib] Issue 21489: Unity3D env lacks group rewards ( #24016 ).
2022-04-21 18:49:52 +02:00
vizdoom_with_attention_net.py
[RLlib] More Trainer -> Algorithm renaming cleanups. ( #25869 )
2022-06-20 15:54:00 +02:00