Skip to content

Commit 4e0ddd9

Browse files
zifeitongtjohnson31415
authored andcommitted
[Bugfix] Properly set distributed_executor_backend in ParallelConfig (vllm-project#4816)
1 parent 6d46185 commit 4e0ddd9

File tree

2 files changed

+8
-3
lines changed

2 files changed

+8
-3
lines changed

vllm/config.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -531,6 +531,7 @@ class ParallelConfig:
531531
If None, will use synchronous tokenization.
532532
ray_workers_use_nsight: Whether to profile Ray workers with nsight, see
533533
https://docs.ray.io/en/latest/ray-observability/user-guides/profiling.html#profiling-nsight-profiler.
534+
placement_group: ray distributed model workers placement group.
534535
distributed_executor_backend: Backend to use for distributed model
535536
workers, either "ray" or "mp" (multiprocessing). If either
536537
pipeline_parallel_size or tensor_parallel_size is greater than 1,

vllm/engine/arg_utils.py

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -548,14 +548,18 @@ def create_engine_config(self, ) -> EngineConfig:
548548
model_config.get_sliding_window(),
549549
self.enable_prefix_caching)
550550
parallel_config = ParallelConfig(
551-
self.pipeline_parallel_size, self.tensor_parallel_size,
552-
self.worker_use_ray, self.max_parallel_loading_workers,
551+
self.pipeline_parallel_size,
552+
self.tensor_parallel_size,
553+
self.worker_use_ray,
554+
self.max_parallel_loading_workers,
553555
self.disable_custom_all_reduce,
554556
TokenizerPoolConfig.create_config(
555557
self.tokenizer_pool_size,
556558
self.tokenizer_pool_type,
557559
self.tokenizer_pool_extra_config,
558-
), self.ray_workers_use_nsight)
560+
),
561+
self.ray_workers_use_nsight,
562+
distributed_executor_backend=self.distributed_executor_backend)
559563

560564
speculative_config = SpeculativeConfig.maybe_create_spec_config(
561565
target_model_config=model_config,

0 commit comments

Comments
 (0)