Commit graph

6227 commits

Author SHA1 Message Date
xwjiang2010
b1496d235f
[tune] fix error handling for fail_fast case. (#22982) 2022-03-10 20:10:05 +00:00
Simon Mo
832354ce3f
[Serve] Compatibility bridge between model wrappers and pipeline (#22995) 2022-03-10 11:52:03 -08:00
qicosmos
e4a9517739
[C++ Worker]Python call cpp worker (#22820) 2022-03-10 11:06:14 -08:00
Yi Cheng
bb5fa6b851
Remove redis in setup.py (#22979) 2022-03-10 11:05:03 -08:00
Archit Kulkarni
c78bd809ce
[job submission] Support local py_modules in jobs (#22843) 2022-03-10 11:42:25 -06:00
Stephanie Wang
85598d9d10
Revert "[ml/tune] Expose new checkpoint interface to users (#22741)" (#23006)
This reverts commit e9692a2a80.
2022-03-10 17:07:44 +00:00
shrekris-anyscale
1100c98222
[serve] Implement Serve Application object (#22917)
The concept of a Serve Application, a data structure containing all information needed to deploy Serve on a Ray cluster, has surfaced during recent design discussions. This change introduces a formal Application data structure and refactors existing code to use it.
2022-03-10 10:28:29 -06:00
Jiajun Yao
2e828cc9e1
Delete dead test_setup_worker.py (#22970)
The tested code is dead so we can remove the code and the test.
2022-03-10 07:20:41 -08:00
Antoni Baum
bf49d37176
[tune] Add Trainable.postprocess_checkpoint (#22973)
Adds postprocess_checkpoint method to Trainable to facilitate the checkpointing of preprocessors in AIR.
2022-03-10 12:14:39 +00:00
Tao Wang
bc14512471
[Hotfix]Fix test_actor failure caused by interface change (#23000) 2022-03-10 19:34:12 +08:00
Kai Fricke
e9692a2a80
[ml/tune] Expose new checkpoint interface to users (#22741)
This PR exposes the new checkpoint interface, implemented in #22691, to end users. It does this by replacing the old external facing TrialCheckpoint class with a merged class that supports the old TrialCheckpoint API (upload, download, save) as well as the new Checkpoint API.

With this PR, users can use the new Checkpoint interface for downstream processing of their Ray Tune results. In a follow-up PR, the new Checkpoint interface will be used internally within Ray Tune and Train for bookkeeping, however, that is not required to unblock the Ray ML use case.
2022-03-10 10:20:24 +00:00
kyle-chen-uber
592656ca28
[horovod] remove deprecated slot concept, use worker instead (#22708)
Horovod updated the attributes of DistributedTrainableCreator and args to create Horovod RayExecutor.
horovod/horovod@a729ba7

The major issue is Horovod deprecated "slot" concept, use "worker" instead, which is more consistent with Generic Ray worker. The issue is currently blocking Uber DL trainers to use raytune.

This commit updates the Horovod RayExecutor init args.

Co-authored-by: Kai Fricke <kai@anyscale.com>
2022-03-10 08:16:42 +00:00
shrekris-anyscale
bc82e2d5c4
[serve] Restore "[serve] Support working_dir in serve run (#22760)" (#22971) 2022-03-09 21:31:23 -08:00
Dmitri Gekhtman
19b4281991
[KubeRay] Pin autoscaler image (#22987)
Sets the autoscaler image to the one from this PR's commit.
#22847
2022-03-09 20:38:37 -08:00
Dmitri Gekhtman
413fe08f87
Move KubeRay autoscaler files into Ray autoscaler directory, add an entry-point. (#22847)
This PR consists of the following clean-up items for KubeRay autoscaler integration:

Remove the docker/kuberay directory

Move the Python files formerly in docker/kuberay to the autoscaler directory.

Use a rayproject/ray image for the autoscaler.

Add an entry point for the kuberay autoscaler to scripts.py. Use the entry point in the example config.

Slightly simplify the code that starts the autoscaler.

Ray versions are updated to Ray 1.11.0, which will be officially released within the next couple of days.

By default, Ray >= 1.11.0 runs without Redis. References to Redis are removed from the example config.

Add the autoscaler configuration test to the CI.

Update development documentation to reflect the changes in this PR.
2022-03-09 18:26:57 -08:00
Jiao
3546aabefd
[7/X][Pipeline] pipeline user facing build function (#22934) 2022-03-09 16:11:11 -08:00
Simon Mo
34ffc7e5cf
[Serve] [3/3 Wrappers] Add Model Wrapper with ray.ml (#22915) 2022-03-09 16:06:59 -08:00
Simon Mo
c844c706bf
[Serve] Use starlette public accessor for Request (#22957) 2022-03-09 13:25:03 -08:00
Jiao
ea9069fef4
[6/X][Pipeline] Add HTTP ingress to serve pipeline (#22878) 2022-03-09 11:39:15 -08:00
Simon Mo
3c4827e0b2
[Serve] [2/3 Wrappers] Add Basic HTTP Adapters (#22914) 2022-03-09 11:36:46 -08:00
Antoni Baum
2ead945438
[datasets] Make label_column optional in to_tf (#22916)
Makes the `label_column` argument in `Dataset.to_tf` optional so that it can be used for prediction.
2022-03-09 11:34:18 -08:00
shrekris-anyscale
61e132b478
[serve] Split test_deploy (#22908)
`test_deploy` has become [flakey](https://flakey-tests.ray.io/#) due to timeout. Since `test_deploy` is already a "large" test, this change splits it into two testing files instead of simply increasing the timeout.
2022-03-09 12:22:51 -06:00
Kai Fricke
b267be4758
[ml] Add Ray ML / AIR checkpoint implementation (#22691)
This PR splits up the changes in #22393 and introduces an implementation of the ML Checkpoint interface used by Ray Tune.

This means, the TuneCheckpoint class implements the to/from_[bytes|dict|directory|object_ref|uri] conversion functions, as well as more high-level functions to transition between the different TuneCheckpoint classes. It also includes test cases for Tune's main conversion modes, i.e. dict - intermediate - dict and fs - intermediate - fs.

These changes will be the basis for refactoring the tune interface to use TuneCheckpoint objects instead of TrialCheckpoints (externally) and instead of paths/objects (internally).
2022-03-09 10:02:59 -08:00
Eric Liang
79a3b56015
[ml] Improve the documentation of ml common classes; add kwargs to predictor (#22936) 2022-03-09 10:01:20 -08:00
Simon Mo
77ead01b65
[Serve] [1/3 Wrappers] Allow @serve.batch to accept args and kwargs (#22913) 2022-03-09 09:15:57 -08:00
Kai Fricke
15601ed79b
Revert "[serve] Support working_dir in serve run (#22760)" (#22956)
This reverts commit ab2741d64b.

The PR breaks ray job submission for anyscale:// URLs
2022-03-09 17:04:46 +00:00
Jiajun Yao
069f5f467c
[Test] Fix and enable test_logging.py (#22904)
Fix and enable test_logging.py
2022-03-09 09:01:38 -08:00
ZhuSenlin
a15890be58
[GCS] refactor the resource related data structures on the GCS (#22924)
* refactor resource data structure in gcs

* fix comment

* fix lint error

* fix

* DISABLED_TestRejectedRequestWorkerLeaseReply as it depends on the update of normal task

Co-authored-by: 黑驰 <senlin.zsl@antgroup.com>
2022-03-09 08:22:02 -08:00
matthewdeng
6b0169b23d
[ml] enable CI tests (#22926)
Follow-up to #22748, enabling tests in CI.

Conditions: A new RAY_CI_ML_AFFECTED condition is added for this test suite. The package currently depends on Ray Data, and will be triggered accordingly.

Dependencies: Adding DATA_PROCESSING_TESTING dependencies (set for install-dependencies.sh) for now.
2022-03-09 14:31:53 +00:00
Jialing He
795b5787dc
[runtime env][bug] Fix RuntimEnv ignore eager_install when _validate is True (#22935)
When _validate is True, RuntimeEnv will ignore field eager_install.
2022-03-09 20:16:55 +08:00
Siyuan (Ryans) Zhuang
b621dc099b
[DAG] Update the example in the doc (#22930)
* update doc
2022-03-08 20:09:45 -08:00
Guyang Song
56287d63e5
[runtime env] remove _rewrite_pip_list_ray_libraries (#22890)
We don't need this logic after using virtualenv.
2022-03-09 11:41:33 +08:00
Stephanie Wang
bf09f5071a
[core] Deflake test_plasma_unlimited (#22911)
test_plasma_unlimited::test_task_unlimited is flaky because one of the assertions is race-y and can trigger after the condition is no longer true (see #22883). This fixes the flake by:
- adding an assertion in between two object allocations to force the object store queue to flush
- keeping one of the ObjectRefs in scope to make sure that the object is still fallback-allocated by the time we reach the failing assertion
2022-03-08 22:00:04 -05:00
Junwen Yao
0395d0987e
[Train] Add support for automatic pipelining of host to device transfer (#22716)
This PR adds the support for concurrently transferring the input from host to device.
2022-03-08 18:37:23 -08:00
Balaji Veeramani
48af260aaf
[Train] Clarify shuffle documentation in prepare_data_loader (#22876)
We essentially use a hack to determine whether shuffling should be enabled in prepare_data_loader. I've clarified the documentation so the hack is easier to understand.
2022-03-08 18:13:29 -08:00
Eric Liang
52491c87e2
Make a pass fixing Dataset API issues (#22886) 2022-03-08 13:07:55 -08:00
shrekris-anyscale
ab2741d64b
[serve] Support working_dir in serve run (#22760)
#22714 added `serve run` to the Serve CLI. This change allows the user to specify a local or remote `working_dir` in `serve run`.
2022-03-08 13:18:41 -06:00
Junwen Yao
d1009c8489
[Train] Add support for metrics aggregation (#22099)
This PR allows users to aggregate metrics returned from all workers.
2022-03-08 11:03:04 -08:00
Balaji Veeramani
37c6169027
[Train] Refactor and add Accelerator classes (#22009)
To support mixed precision (see #20643), we need to store a GradScaler instance that is accessibly by both prepare_optimizer and backward functions (these functions will be added later).

This PR introduces the Accelerator, an object that implements methods to perform backend-specific training optimizations.
2022-03-08 10:26:00 -08:00
Balaji Veeramani
04b10ff9e9
[Train] Tell user to specify cluster address if placement group times out (#22845)
If you don't add `ray.init("auto")` to your training script, then your training script might complain that there aren't enough resources, even if `ray status` shows that there are.

Co-authored-by: Amog Kamsetty <amogkam@users.noreply.github.com>
2022-03-08 10:24:12 -08:00
matthewdeng
7b5813e94f
[ml] add initial Dataset Preprocessors (#22748) 2022-03-08 09:59:03 -08:00
Gagandeep Singh
2899dc1bb5
Fixed MRO for DerivedActorClass (#22113)
Comments to be noted from the discussion below,

https://github.com/ray-project/ray/pull/22113#discussion_r802512907

> Problem - We cannot always delegate call to cls.__init__ or modified_cls.__init__. Because if always delegate call to cls.__init__ from here, then user defined class's __init__ method will be ignore leading to issues like, https://github.com/ray-project/ray/issues/21868. If we always delegate call to modified_cls.__init__ then it will allow inheriting from actor classes leading to failure of test_actor_inheritance. So, I have added this if-else check to figure out which __init__ method should be called. If "__module__", "__qualname__" and "__init__" are present in args[-1] then it would mean an actor class is being inherited so cls.__init__ should be called. However, if no such signal is received in args then user defined class's __init__ i.e., modified_class.__init__ should be called.

https://github.com/ray-project/ray/pull/22113#discussion_r808696261

> So I noted that ActorClass.__init__ will anyway raise a TypeError whenever it will be inherited. To exactly figure out whether the exception is due to inheritance of ActorClass, I created a new class ActorClassInheritanceException(TypeError). Now, whenever this will be raised, then DerivedActorClass will get a clear signal about inheritance of ActorClass. In other cases, it will be safe to conclude (AFAICT) that user called __init__ method of their class and we will proceed normally. IMHO, this is a better and more robust solution which just depends on a simple signal i.e., raising a particular exception in a specific event. It doesn't matter how inheritance is prevented as in the end we just need to raise ActorClassInheritanceException and all other code will be able to detect that easily.

https://github.com/ray-project/ray/pull/22113#issuecomment-1048527387
2022-03-08 09:37:19 -08:00
xwjiang2010
f5995dccdf
[tune] Trainables will now know TUNE_ORIG_WORKING_DIR (#22803)
Also updated the docs.
2022-03-08 15:56:30 +00:00
Jiajun Yao
7f57268bd0
Fix duplidate test bazel target (#22892) 2022-03-08 14:29:13 +09:00
Jiajun Yao
4801e57c77
[Test] Add missing tests to bazel BUILD (#22827) 2022-03-07 19:54:49 -08:00
Jian Xiao
c2908de401
For a dataset comprised of both empty and non-empty blocks, let the non-empty blocks determine the schema (#22834)
There is a bug in combining the results from map_batches: if we create two dataset out of the same data, but with different num of partitions, we may get different results when run the same map_batches() on them. That is, num of partitions is affecting the map_batches() results, which should not.
2022-03-07 18:17:49 -08:00
Jiajun Yao
2302b4eea8
Stop and join actor asyncio threads during exit (#22810) 2022-03-07 14:45:08 -08:00
Stephanie Wang
fa14120f93
Move tests out of test_object_spilling to de-flake (#22831)
This test is timing out often in debug_mode, so moved some tests to test_object_spilling_3.
2022-03-07 17:39:55 -05:00
SangBin Cho
79e8405fda
Revert "[GCS] refactor the resource related data structures on the GCS (#22817)" (#22863)
This reverts commit 549466a42f.
2022-03-07 08:48:17 -08:00
shrekris-anyscale
15d97a1021
[serve] Support init_args and init_kwargs in serve run (#22805)
#22714 added `serve run` to the Serve CLI. This change allows the user to specify `init_args` and `init_kwargs` in `serve run` if they are deploying via import path.
2022-03-07 09:45:17 -06:00