ray/release/air_tests/air_benchmarks/workloads
Kai Fricke cf75cf7232
[air] Add AIR distributed training benchmark for Torch FashionMNIST (#26436)
This PR adds a distributed benchmark test for Pytorch MNIST training. It compares training with Ray AIR with training with vanilla PyTorch.

In both cases, the same training loop is used. For Ray AIR, we use a TorchTrainer with 4 CPU workers. For vanilla PyTorch, we upload a training script and kick it off (using Ray tasks) in subprocesses on each node. In both cases, we collect the end to end runtime.

Signed-off-by: Kai Fricke <kai@anyscale.com>
2022-07-13 10:53:24 +01:00
..
_torch_prepare.py [air] Add AIR distributed training benchmark for Torch FashionMNIST (#26436) 2022-07-13 10:53:24 +01:00
benchmark_util.py [air] Add AIR distributed training benchmark for Torch FashionMNIST (#26436) 2022-07-13 10:53:24 +01:00
torch_benchmark.py [air] Add AIR distributed training benchmark for Torch FashionMNIST (#26436) 2022-07-13 10:53:24 +01:00