ray/examples/carla
Eric Liang 6724f57b03 [Examples] Add Carla test env (#1343)
* add carla example

* add reward

* set obs

* Sun Dec 17 16:06:00 PST 2017

* add spec

* fix measurement

* add train script

* resize to 80x80

* null

* initial small training run

* robustify env, clean up action space

* clean up vars

* switch to town2 which is faster

* tunify train.py

* add discrete mode

* update

* fix excessive brakinG

* fix the weather

* rename

* redirect output and from future import

* doc

* update

* fix rebase

* allow dqn gpu growht

* adjust dqn hyperparams

* better ppo parameters
2017-12-19 12:57:58 -08:00
..
env.py [Examples] Add Carla test env (#1343) 2017-12-19 12:57:58 -08:00
README [Examples] Add Carla test env (#1343) 2017-12-19 12:57:58 -08:00
train_dqn.py [Examples] Add Carla test env (#1343) 2017-12-19 12:57:58 -08:00
train_ppo.py [Examples] Add Carla test env (#1343) 2017-12-19 12:57:58 -08:00

(Experimental) gym environment for https://github.com/carla-simulator/carla

To run, first download and unpack the Carla release from this URL: https://github.com/carla-simulator/carla/releases/tag/0.7.0

Then, you can try running env.py to drive the car. Run train_ppo.py or train_dqn.py to attempt training.

    $ pkill -9 Carla
    $ export PYTHONPATH=/home/ubuntu/CARLA_0.7.0/PythonClient:$PYTHONPATH
    $ export CARLA_SERVER=/home/ubuntu/CARLA_0.7.0/CarlaUE4.sh
    $ python env.py

Note that the reward function is currently hard-coded to drive straight down the street.