mirror of
https://github.com/vale981/ray
synced 2025-03-06 10:31:39 -05:00

* Add XGBoost release tests * Add more xgboost release tests * Use failure state manager * Add release test documentation * Fix wording * Automate fault tolerance tests
113 lines
3.5 KiB
Python
113 lines
3.5 KiB
Python
"""Fault tolerance test (small cluster, non-elastic training)
|
|
|
|
In this run, two training actors will die after some time. It is expected that
|
|
in both cases xgboost_ray stops training, restarts the dead actors, and
|
|
continues training with all four actors.
|
|
|
|
Test owner: krfricke
|
|
|
|
Acceptance criteria: Should run through and report final results. Intermediate
|
|
output should show that training halts wenn an actor dies and continues only
|
|
when all four actors are available again. The test will fail if fault
|
|
tolerance did not work correctly.
|
|
|
|
Notes: This test seems to be somewhat flaky. This might be due to
|
|
race conditions in handling dead actors. This is likely a problem of
|
|
the xgboost_ray implementation and not of this test.
|
|
"""
|
|
import os
|
|
import time
|
|
|
|
import ray
|
|
from ray.services import get_node_ip_address
|
|
|
|
from xgboost_ray import RayParams
|
|
from xgboost_ray.session import get_actor_rank, put_queue
|
|
|
|
from xgboost.callback import TrainingCallback
|
|
from xgboost.rabit import get_world_size
|
|
|
|
from _train import train_ray
|
|
|
|
|
|
@ray.remote
|
|
class FailureState:
|
|
def __init__(self):
|
|
self._failed_ids = set()
|
|
|
|
def set_failed(self, id):
|
|
if id in self._failed_ids:
|
|
return False
|
|
self._failed_ids.add(id)
|
|
return True
|
|
|
|
def has_failed(self, id):
|
|
return id in self._failed_ids
|
|
|
|
|
|
class FailureInjection(TrainingCallback):
|
|
def __init__(self, id, state, ranks, iteration, allow_ips=None):
|
|
self._id = id
|
|
self._state = state
|
|
self._ranks = ranks or []
|
|
self._iteration = iteration
|
|
self._allow_ips = allow_ips
|
|
super(FailureInjection).__init__()
|
|
|
|
def after_iteration(self, model, epoch, evals_log):
|
|
if self._allow_ips and get_node_ip_address() not in self._allow_ips:
|
|
return
|
|
|
|
if epoch == self._iteration:
|
|
rank = get_actor_rank()
|
|
if rank in self._ranks:
|
|
if not ray.get(self._state.has_failed.remote(id)):
|
|
success = ray.get(self._state.set_failed.remote(id))
|
|
if not success:
|
|
# Another rank is already about to fail
|
|
return
|
|
|
|
pid = os.getpid()
|
|
print(f"Killing process: {pid} for actor rank {rank}")
|
|
time.sleep(1)
|
|
os.kill(pid, 9)
|
|
|
|
|
|
class TrackingCallback(TrainingCallback):
|
|
def after_iteration(self, model, epoch, evals_log):
|
|
put_queue(get_world_size())
|
|
|
|
|
|
if __name__ == "__main__":
|
|
ray.init(address="auto")
|
|
|
|
failure_state = FailureState.remote()
|
|
|
|
ray_params = RayParams(
|
|
elastic_training=False,
|
|
max_actor_restarts=2,
|
|
num_actors=4,
|
|
cpus_per_actor=4,
|
|
gpus_per_actor=0)
|
|
|
|
_, additional_results, _ = train_ray(
|
|
path="/data/classification.parquet",
|
|
num_workers=4,
|
|
num_boost_rounds=100,
|
|
num_files=200,
|
|
regression=False,
|
|
use_gpu=False,
|
|
ray_params=ray_params,
|
|
xgboost_params=None,
|
|
callbacks=[
|
|
TrackingCallback(),
|
|
FailureInjection(
|
|
id="first_fail", state=failure_state, ranks=[2], iteration=14),
|
|
FailureInjection(
|
|
id="second_fail", state=failure_state, ranks=[0], iteration=34)
|
|
])
|
|
|
|
actor_1_world_size = set(additional_results["callback_returns"][1])
|
|
assert len(actor_1_world_size) == 1 and 4 in actor_1_world_size, \
|
|
"Training with fewer than 4 actors observed, but this was " \
|
|
"non-elastic training. Please report to test owner."
|