Skip to content

Commit 7d72664

Browse files
committed
[Frontend] remove --device arg
Signed-off-by: Kebe <[email protected]>
1 parent 1b1e8e0 commit 7d72664

File tree

2 files changed

+2
-12
lines changed

2 files changed

+2
-12
lines changed

docs/source/getting_started/installation/gpu/xpu.inc.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -66,7 +66,6 @@ XPU platform supports **tensor parallel** inference/serving and also supports **
6666
python -m vllm.entrypoints.openai.api_server \
6767
--model=facebook/opt-13b \
6868
--dtype=bfloat16 \
69-
--device=xpu \
7069
--max_model_len=1024 \
7170
--distributed-executor-backend=ray \
7271
--pipeline-parallel-size=2 \

vllm/engine/arg_utils.py

Lines changed: 2 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@
1919
import vllm.envs as envs
2020
from vllm.config import (BlockSize, CacheConfig, CacheDType, CompilationConfig,
2121
ConfigFormat, ConfigType, DecodingConfig,
22-
DetailedTraceModules, Device, DeviceConfig,
22+
DetailedTraceModules, DeviceConfig,
2323
DistributedExecutorBackend, GuidedDecodingBackend,
2424
GuidedDecodingBackendV1, HfOverrides, KVEventsConfig,
2525
KVTransferConfig, LoadConfig, LoadFormat, LoRAConfig,
@@ -354,7 +354,6 @@ class EngineArgs:
354354
max_prompt_adapter_token: int = \
355355
PromptAdapterConfig.max_prompt_adapter_token
356356

357-
device: Device = DeviceConfig.device
358357
num_scheduler_steps: int = SchedulerConfig.num_scheduler_steps
359358
multi_step_stream_outputs: bool = SchedulerConfig.multi_step_stream_outputs
360359
ray_workers_use_nsight: bool = ParallelConfig.ray_workers_use_nsight
@@ -731,14 +730,6 @@ def add_cli_args(parser: FlexibleArgumentParser) -> FlexibleArgumentParser:
731730
"--max-prompt-adapter-token",
732731
**prompt_adapter_kwargs["max_prompt_adapter_token"])
733732

734-
# Device arguments
735-
device_kwargs = get_kwargs(DeviceConfig)
736-
device_group = parser.add_argument_group(
737-
title="DeviceConfig",
738-
description=DeviceConfig.__doc__,
739-
)
740-
device_group.add_argument("--device", **device_kwargs["device"])
741-
742733
# Speculative arguments
743734
speculative_group = parser.add_argument_group(
744735
title="SpeculativeConfig",
@@ -977,7 +968,7 @@ def create_engine_config(
977968
from vllm.platforms import current_platform
978969
current_platform.pre_register_and_update()
979970

980-
device_config = DeviceConfig(device=self.device)
971+
device_config = DeviceConfig(device=current_platform.device_type)
981972
model_config = self.create_model_config()
982973

983974
# * If VLLM_USE_V1 is unset, we enable V1 for "supported features"

0 commit comments

Comments
 (0)