[BugFix] Avoid unnecessary Ray import warnings (#6079)
This commit is contained in:
parent
d18bab3587
commit
d830656a97
@ -682,11 +682,13 @@ class ParallelConfig:
|
|||||||
|
|
||||||
from vllm.executor import ray_utils
|
from vllm.executor import ray_utils
|
||||||
backend = "mp"
|
backend = "mp"
|
||||||
ray_found = ray_utils.ray is not None
|
ray_found = ray_utils.ray_is_available()
|
||||||
if cuda_device_count_stateless() < self.world_size:
|
if cuda_device_count_stateless() < self.world_size:
|
||||||
if not ray_found:
|
if not ray_found:
|
||||||
raise ValueError("Unable to load Ray which is "
|
raise ValueError("Unable to load Ray which is "
|
||||||
"required for multi-node inference")
|
"required for multi-node inference, "
|
||||||
|
"please install Ray with `pip install "
|
||||||
|
"ray`.") from ray_utils.ray_import_err
|
||||||
backend = "ray"
|
backend = "ray"
|
||||||
elif ray_found:
|
elif ray_found:
|
||||||
if self.placement_group:
|
if self.placement_group:
|
||||||
@ -718,6 +720,9 @@ class ParallelConfig:
|
|||||||
raise ValueError(
|
raise ValueError(
|
||||||
"Unrecognized distributed executor backend. Supported values "
|
"Unrecognized distributed executor backend. Supported values "
|
||||||
"are 'ray' or 'mp'.")
|
"are 'ray' or 'mp'.")
|
||||||
|
if self.distributed_executor_backend == "ray":
|
||||||
|
from vllm.executor import ray_utils
|
||||||
|
ray_utils.assert_ray_available()
|
||||||
if not self.disable_custom_all_reduce and self.world_size > 1:
|
if not self.disable_custom_all_reduce and self.world_size > 1:
|
||||||
if is_hip():
|
if is_hip():
|
||||||
self.disable_custom_all_reduce = True
|
self.disable_custom_all_reduce = True
|
||||||
|
|||||||
@ -380,6 +380,11 @@ class AsyncLLMEngine:
|
|||||||
"""Creates an async LLM engine from the engine arguments."""
|
"""Creates an async LLM engine from the engine arguments."""
|
||||||
# Create the engine configs.
|
# Create the engine configs.
|
||||||
engine_config = engine_args.create_engine_config()
|
engine_config = engine_args.create_engine_config()
|
||||||
|
|
||||||
|
if engine_args.engine_use_ray:
|
||||||
|
from vllm.executor import ray_utils
|
||||||
|
ray_utils.assert_ray_available()
|
||||||
|
|
||||||
distributed_executor_backend = (
|
distributed_executor_backend = (
|
||||||
engine_config.parallel_config.distributed_executor_backend)
|
engine_config.parallel_config.distributed_executor_backend)
|
||||||
|
|
||||||
|
|||||||
@ -42,14 +42,26 @@ try:
|
|||||||
output = pickle.dumps(output)
|
output = pickle.dumps(output)
|
||||||
return output
|
return output
|
||||||
|
|
||||||
|
ray_import_err = None
|
||||||
|
|
||||||
except ImportError as e:
|
except ImportError as e:
|
||||||
logger.warning(
|
|
||||||
"Failed to import Ray with %r. For multi-node inference, "
|
|
||||||
"please install Ray with `pip install ray`.", e)
|
|
||||||
ray = None # type: ignore
|
ray = None # type: ignore
|
||||||
|
ray_import_err = e
|
||||||
RayWorkerWrapper = None # type: ignore
|
RayWorkerWrapper = None # type: ignore
|
||||||
|
|
||||||
|
|
||||||
|
def ray_is_available() -> bool:
|
||||||
|
"""Returns True if Ray is available."""
|
||||||
|
return ray is not None
|
||||||
|
|
||||||
|
|
||||||
|
def assert_ray_available():
|
||||||
|
"""Raise an exception if Ray is not available."""
|
||||||
|
if ray is None:
|
||||||
|
raise ValueError("Failed to import Ray, please install Ray with "
|
||||||
|
"`pip install ray`.") from ray_import_err
|
||||||
|
|
||||||
|
|
||||||
def initialize_ray_cluster(
|
def initialize_ray_cluster(
|
||||||
parallel_config: ParallelConfig,
|
parallel_config: ParallelConfig,
|
||||||
ray_address: Optional[str] = None,
|
ray_address: Optional[str] = None,
|
||||||
@ -65,10 +77,7 @@ def initialize_ray_cluster(
|
|||||||
ray_address: The address of the Ray cluster. If None, uses
|
ray_address: The address of the Ray cluster. If None, uses
|
||||||
the default Ray cluster address.
|
the default Ray cluster address.
|
||||||
"""
|
"""
|
||||||
if ray is None:
|
assert_ray_available()
|
||||||
raise ImportError(
|
|
||||||
"Ray is not installed. Please install Ray to use multi-node "
|
|
||||||
"serving.")
|
|
||||||
|
|
||||||
# Connect to a ray cluster.
|
# Connect to a ray cluster.
|
||||||
if is_hip() or is_xpu():
|
if is_hip() or is_xpu():
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user