ray/rllib/examples
2020-10-01 16:57:10 +02:00
..
env [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
export [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
models [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
policy [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
serving [RLlib] Deprecate old classes, methods, functions, config keys (in prep for RLlib 1.0). (#10544) 2020-09-06 10:58:00 +02:00
__init__.py [rllib] Try moving RLlib to top level dir (#5324) 2019-08-05 23:25:49 -07:00
attention_net.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
attention_net_supervised.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
autoregressive_action_dist.py [RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520) 2020-05-27 16:19:13 +02:00
batch_norm_model.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
cartpole_lstm.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
centralized_critic.py [RLlib] PPO, APPO, and DD-PPO code cleanup. (#10420) 2020-09-02 14:03:01 +02:00
centralized_critic_2.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
complex_struct_space.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
custom_env.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
custom_eval.py [RLlib] SAC algo cleanup. (#10825) 2020-09-20 11:27:02 +02:00
custom_fast_model.py [RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520) 2020-05-27 16:19:13 +02:00
custom_keras_model.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
custom_loss.py [rllib] Remove extra model config kwargs passed incorrectly for Torch models (#10055) 2020-08-17 11:12:20 -07:00
custom_metrics_and_callbacks.py [tune] API revamp fix (#10518) 2020-09-05 15:34:53 -07:00
custom_metrics_and_callbacks_legacy.py [tune] API revamp fix (#10518) 2020-09-05 15:34:53 -07:00
custom_rnn_model.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
custom_tf_policy.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
custom_torch_policy.py [RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520) 2020-05-27 16:19:13 +02:00
custom_train_fn.py [RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520) 2020-05-27 16:19:13 +02:00
dmlab_watermaze.py Remove future imports (#6724) 2020-01-09 00:15:48 -08:00
eager_execution.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
hierarchical_training.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
mobilenet_v2_with_lstm.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
multi_agent_cartpole.py [RLlib] Tf2x preparation; part 2 (upgrading try_import_tf()). (#9136) 2020-06-30 10:13:20 +02:00
multi_agent_custom_policy.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
multi_agent_independent_learning.py Multi-agent Algorithm Documentation Updates (#9722) 2020-09-03 22:37:46 -07:00
multi_agent_parameter_sharing.py [RLlib] Deprecate old classes, methods, functions, config keys (in prep for RLlib 1.0). (#10544) 2020-09-06 10:58:00 +02:00
multi_agent_two_trainers.py [RLlib] Auto-framework, retire use_pytorch in favor of framework=... (#8520) 2020-05-27 16:19:13 +02:00
nested_action_spaces.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
parametric_actions_cartpole.py [RLlib] Issue #9366 (DQN w/o dueling produces invalid actions). (#9386) 2020-07-10 12:43:03 +02:00
pettingzoo_env.py Fixed PettingZooEnv (#10847) 2020-09-17 11:28:42 -07:00
rock_paper_scissors_multiagent.py [RLlib] Trajectory view API: Simple List Collector (on by default for PPO); LSTM-agnostic (#11056) 2020-10-01 16:57:10 +02:00
rollout_worker_custom_workflow.py ci: Redo format.sh --all script & backfill lint fixes (#9956) 2020-08-07 16:49:49 -07:00
saving_experiences.py [rllib] Add execution module to package ref (#10941) 2020-09-21 23:03:06 -07:00
two_step_game.py [RLlib] Issue 8384: QMIX doesn't learn anything. (#9527) 2020-07-17 12:14:34 +02:00
two_trainer_workflow.py [RLlib] Deprecate old classes, methods, functions, config keys (in prep for RLlib 1.0). (#10544) 2020-09-06 10:58:00 +02:00
unity3d_env_local.py [RLlib] Unity blogpost final fixes. (#10894) 2020-09-20 14:13:20 +02:00