mirror of
https://github.com/vale981/ray
synced 2025-03-09 21:06:39 -04:00
![]() This adds a simple DQN+PPO example for multi-agent. We don't do anything fancy here, just syncing weights between two separate trainers. This potentially is wasting some compute, but is very simple to set up. It might be nice to share experience collection between the top-level trainers in the future. |
||
---|---|---|
.. | ||
multi_node_tests | ||
multi_node_docker_test.py | ||
run_asv.sh | ||
run_multi_node_tests.sh | ||
run_rllib_asv.sh |