Artur Niederfahrenhorst
|
0dceddb912
|
[RLlib] Move learning_starts logic from buffers into training_step() . (#26032)
|
2022-08-11 13:07:30 +02:00 |
|
Rohan Potdar
|
09ce4711fd
|
[RLlib]: Move OPE to evaluation config (#25911)
|
2022-07-12 11:04:34 -07:00 |
|
Avnish Narayan
|
1243ed62bf
|
[RLlib] Make Dataset reader default reader and enable CRR to use dataset (#26304)
Co-authored-by: avnish <avnish@avnishs-MBP.local.meter>
|
2022-07-08 12:43:35 -07:00 |
|
Rohan Potdar
|
a9d8da0100
|
[RLlib]: Doubly Robust Off-Policy Evaluation. (#25056)
|
2022-06-07 12:52:19 +02:00 |
|
Sven Mika
|
b5bc2b93c3
|
[RLlib] Move all remaining algos into algorithms directory. (#25366)
|
2022-06-04 07:35:24 +02:00 |
|
Rohan Potdar
|
ab81c8e9ca
|
[RLlib]: Rename input_evaluation to off_policy_estimation_methods . (#25107)
|
2022-05-27 13:14:54 +02:00 |
|
Artur Niederfahrenhorst
|
d76ef9add5
|
[RLLib] Fix RNNSAC example failing on CI + fixes for recurrent models for other Q Learning Algos. (#24923)
|
2022-05-24 14:39:43 +02:00 |
|
Steven Morad
|
501d932449
|
[RLlib] SAC, RNNSAC, and CQL TrainerConfig objects (#25059)
|
2022-05-22 19:58:47 +02:00 |
|
Jun Gong
|
68a9a33386
|
[RLlib] Retry agents -> algorithms. with proper doc changes this time. (#24797)
|
2022-05-16 09:45:32 +02:00 |
|
Simon Mo
|
9f23affdc0
|
[Hotfix] Unbreak lint in master (#24794)
|
2022-05-13 15:05:05 -07:00 |
|
kourosh hakhamaneshi
|
ffcbb30552
|
[RLlib] Move from agents to algorithms - CQL, MARWIL, AlphaStar, MAML, Dreamer, MBMPO. (#24739)
|
2022-05-13 18:43:36 +02:00 |
|