ray/release/xgboost_tests
2021-08-16 12:44:25 +02:00
..
workloads Create a wait_for_num_nodes() function, and use it in train_small (#16784) 2021-07-01 10:17:53 +01:00
app_config.yaml [release] Fix app config: Install latest releases. Bump xgboost-ray version (#16581) 2021-06-24 12:56:21 +01:00
app_config_gpu.yaml [release] update/unify base images (#17859) 2021-08-16 12:44:25 +02:00
cluster_cpu_moderate.yaml [release] Move xgboost tune small + microbenchmark release test to new release automation (#15619) 2021-05-08 20:38:39 +01:00
cluster_cpu_small.yaml [release] Move xgboost tune small + microbenchmark release test to new release automation (#15619) 2021-05-08 20:38:39 +01:00
cluster_gpu_small.yaml [release] Move xgboost tune small + microbenchmark release test to new release automation (#15619) 2021-05-08 20:38:39 +01:00
create_test_data.py [xgboost] Add XGBoost release tests (#13456) 2021-01-20 18:40:23 +01:00
README.rst [xgboost] Update XGBoost release test configs (#13941) 2021-02-17 23:00:49 +01:00
requirements.txt [xgboost] Add XGBoost release tests (#13456) 2021-01-20 18:40:23 +01:00
setup_xgboost.sh [release] release 1.3.0 results and test updates (#15366) 2021-05-04 22:10:04 +01:00
tpl_cpu_moderate.yaml [release] Move xgboost tune small + microbenchmark release test to new release automation (#15619) 2021-05-08 20:38:39 +01:00
tpl_cpu_small.yaml [release] Move xgboost tune small + microbenchmark release test to new release automation (#15619) 2021-05-08 20:38:39 +01:00
tpl_gpu_small.yaml [release] Move xgboost tune small + microbenchmark release test to new release automation (#15619) 2021-05-08 20:38:39 +01:00
wait_cluster.py [xgboost] Add XGBoost release tests (#13456) 2021-01-20 18:40:23 +01:00
xgboost_tests.yaml Roll forward to run train_small in client mode. (#16610) 2021-06-23 08:52:08 +01:00

XGBoost on Ray tests
====================

This directory contains various XGBoost on Ray release tests.

You should run these tests with the `releaser <https://github.com/ray-project/releaser>`_ tool.

Overview
--------
There are four kinds of tests:

1. ``distributed_api_test`` - checks general API functionality and should finish very quickly (< 1 minute)
2. ``train_*`` - checks single trial training on different setups.
3. ``tune_*`` - checks multi trial training via Ray Tune.
4. ``ft_*`` - checks fault tolerance.

Generally the releaser tool will run all tests in parallel, but if you do
it sequentially, be sure to do it in the order above. If ``train_*`` fails,
``tune_*`` will fail, too.

Acceptance criteria
-------------------
These tests are considered passing when they throw no error at the end of
the output log.