ray/benchmarks/object_store/test_object_store.py
SangBin Cho 97b4490401
[Nightly Test] Readjust nightly test schedule (#20717)
- Removing scale_to logic from object store. We don't need to scale during tests, which will disambiguate infra failures vs app failures.
- Run microbenchmark in core nightly, meaning it will run even more often
- Run weekly scalability tests daily instead. (They are not too expensive).
- Run some core daily tests separately to avoid infra failures.
2021-11-26 06:59:16 -08:00

65 lines
1.4 KiB
Python

import numpy as np
import ray
import ray.autoscaler.sdk
import json
import os
from time import perf_counter
from tqdm import tqdm
NUM_NODES = 50
OBJECT_SIZE = 2**30
def num_alive_nodes():
n = 0
for node in ray.nodes():
if node["Alive"]:
n += 1
return n
def test_object_broadcast():
assert num_alive_nodes() == NUM_NODES
@ray.remote(num_cpus=1, resources={"node": 1})
class Actor:
def foo(self):
pass
def sum(self, arr):
return np.sum(arr)
actors = [Actor.remote() for _ in range(NUM_NODES)]
arr = np.ones(OBJECT_SIZE, dtype=np.uint8)
ref = ray.put(arr)
for actor in tqdm(actors, desc="Ensure all actors have started."):
ray.get(actor.foo.remote())
result_refs = []
for actor in tqdm(actors, desc="Broadcasting objects"):
result_refs.append(actor.sum.remote(ref))
results = ray.get(result_refs)
for result in results:
assert result == OBJECT_SIZE
ray.init(address="auto")
start = perf_counter()
test_object_broadcast()
end = perf_counter()
print(f"Broadcast time: {end - start} ({OBJECT_SIZE} B x {NUM_NODES} nodes)")
if "TEST_OUTPUT_JSON" in os.environ:
out_file = open(os.environ["TEST_OUTPUT_JSON"], "w")
results = {
"broadcast_time": end - start,
"object_size": OBJECT_SIZE,
"num_nodes": NUM_NODES,
"success": "1"
}
json.dump(results, out_file)