.. |
exploration
|
[RLlib] Fix "Cannot convert a symbolic Tensor (default_policy/strided_slice_3:0) to a numpy array. This error may indicate that you're trying to pass a Tensor to a NumPy call, which is not supported" (#17587)
|
2021-08-05 11:39:15 -04:00 |
schedules
|
[RLlib] Allow for more than 2^31 policy timesteps. (#11301)
|
2020-10-12 13:49:11 -07:00 |
spaces
|
[RLlib] Redo fix bug normalize vs unsquash actions (original PR made log-likelihood test flakey). (#17014)
|
2021-07-13 14:01:30 -04:00 |
tests
|
Change Python's ObjectID to ObjectRef (#9353)
|
2020-07-10 17:49:04 +08:00 |
__init__.py
|
[RLlib] Add @Deprecated decorator to simplify/unify deprecation of classes, methods, functions. (#17530)
|
2021-08-03 18:30:02 -04:00 |
actors.py
|
Change Python's ObjectID to ObjectRef (#9353)
|
2020-07-10 17:49:04 +08:00 |
annotations.py
|
[rllib] Fix classes decorated with @Deprecated to be classes instead of methods (#17666)
|
2021-08-10 18:25:31 -07:00 |
compression.py
|
Stop vendoring pyarrow (#7233)
|
2020-02-19 19:01:26 -08:00 |
debug.py
|
[rllib] Flexible multi-agent replay modes and replay_sequence_length (#8893)
|
2020-06-12 20:17:27 -07:00 |
deprecation.py
|
[RLlib] Add @Deprecated decorator to simplify/unify deprecation of classes, methods, functions. (#17530)
|
2021-08-03 18:30:02 -04:00 |
error.py
|
Remove future imports (#6724)
|
2020-01-09 00:15:48 -08:00 |
filter.py
|
[rllib] Rrk/12079 custom filters (#12095)
|
2020-11-19 13:20:20 -08:00 |
filter_manager.py
|
[rllib] Deprecate policy optimizers (#8345)
|
2020-05-21 10:16:18 -07:00 |
framework.py
|
[RLlib] Add @Deprecated decorator to simplify/unify deprecation of classes, methods, functions. (#17530)
|
2021-08-03 18:30:02 -04:00 |
from_config.py
|
[RLlib] Make envs specifiable in configs by their class path. (#8750)
|
2020-06-03 08:14:29 +02:00 |
images.py
|
[RLlib] CV2 to Skimage dependency change (#16841)
|
2021-07-21 22:24:18 -04:00 |
memory.py
|
[RLlib] Trajectory View API (preparatory cleanup and enhancements). (#9678)
|
2020-07-29 21:15:09 +02:00 |
numpy.py
|
[RLlib] Policies get/set_state fixes and enhancements. (#16354)
|
2021-06-15 13:08:43 +02:00 |
sgd.py
|
[RLlib] Redo simplify multi agent config dict: Reverted b/c seemed to break test_typing (non RLlib test). (#17046)
|
2021-07-15 05:51:24 -04:00 |
test_utils.py
|
[rllib] Improve test learning check, fix flaky two step qmix (#16843)
|
2021-07-06 19:39:12 +01:00 |
tf_ops.py
|
[RLlib] Issues: 17397, 17425, 16715, 17174. When on driver, Torch|TFPolicy should not use ray.get_gpu_ids() (b/c no GPUs assigned by ray). (#17444)
|
2021-08-02 17:29:59 -04:00 |
tf_run_builder.py
|
[RLlib] Tf2x preparation; part 2 (upgrading try_import_tf() ). (#9136)
|
2020-06-30 10:13:20 +02:00 |
threading.py
|
[RLlib] R2D2 Implementation. (#13933)
|
2021-02-25 12:18:11 +01:00 |
timer.py
|
[rllib] Enable performance metrics reporting for RLlib pipelines, add A3C (#7299)
|
2020-02-28 16:44:17 -08:00 |
torch_ops.py
|
[RLlib] Torch algos use now-framework-agnostic MultiGPUTrainOneStep execution op (~33% speedup for PPO-torch + GPU). (#17371)
|
2021-08-03 11:35:49 -04:00 |
typing.py
|
[RLlib] Redo simplify multi agent config dict: Reverted b/c seemed to break test_typing (non RLlib test). (#17046)
|
2021-07-15 05:51:24 -04:00 |
window_stat.py
|
[RLLib] WindowStat bug fix (#9213)
|
2020-07-12 23:01:32 +02:00 |