Skip to content

Commit 31074c2

Browse files
committed
fix: make vllm --version work
1 parent 292aec2 commit 31074c2

File tree

4 files changed

+36
-33
lines changed

4 files changed

+36
-33
lines changed

vllm/config.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -28,8 +28,7 @@
2828
import vllm.envs as envs
2929
from vllm.compilation.inductor_pass import CallableInductorPass, InductorPass
3030
from vllm.logger import init_logger
31-
from vllm.model_executor.layers.quantization import (QUANTIZATION_METHODS,
32-
get_quantization_config)
31+
from vllm.model_executor.layers.quantization import QUANTIZATION_METHODS
3332
from vllm.platforms import CpuArchEnum
3433
from vllm.sampling_params import GuidedDecodingParams
3534
from vllm.tracing import is_otel_available, otel_import_error_traceback
@@ -42,8 +41,8 @@
4241
from vllm.transformers_utils.utils import is_s3, maybe_model_redirect
4342
from vllm.utils import (GiB_bytes, LayerBlockType, LazyLoader,
4443
cuda_device_count_stateless, get_cpu_memory,
45-
get_open_port, is_torch_equal_or_newer,
46-
random_uuid, resolve_obj_by_qualname)
44+
get_open_port, is_torch_equal_or_newer, random_uuid,
45+
resolve_obj_by_qualname)
4746

4847
if TYPE_CHECKING:
4948
from _typeshed import DataclassInstance
@@ -68,7 +67,6 @@
6867
HfOverrides = None
6968
QuantizationConfig = None
7069

71-
from packaging.version import Version
7270

7371
logger = init_logger(__name__)
7472

vllm/engine/arg_utils.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -291,7 +291,9 @@ def can_be_type(cls: TypeHint, type: TypeHintT) -> TypeIs[TypeHintT]:
291291

292292
def is_custom_type(cls: TypeHint) -> bool:
293293
"""Check if the class is a custom type."""
294-
return cls.__module__ != "builtins"
294+
if isinstance(cls, type):
295+
return cls.__module__ != "builtins"
296+
return True
295297

296298
def get_kwargs(cls: type[Any]) -> dict[str, Any]:
297299
cls_docs = get_attr_docs(cls)

vllm/entrypoints/chat_utils.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,7 @@
11
# SPDX-License-Identifier: Apache-2.0
22

3+
from __future__ import annotations
4+
35
import asyncio
46
import json
57
from abc import ABC, abstractmethod

vllm/transformers_utils/config.py

Lines changed: 28 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -28,39 +28,40 @@
2828
# yapf: disable
2929
if TYPE_CHECKING:
3030
import huggingface_hub as hfhub
31-
import huggingface_hub.utils as hfhub_utils
32-
from transformers import GenerationConfig, PretrainedConfig
31+
import huggingface_hub.errors as hfhub_errors
32+
from transformers.configuration_utils import PretrainedConfig
33+
from transformers.generation.configuration_utils import GenerationConfig
3334
else:
3435
hfhub = LazyLoader("hfhub", globals(), "huggingface_hub")
35-
hfhub_utils = LazyLoader("hfhub_utils", globals(), "huggingface_hub.utils")
36+
hfhub_errors = LazyLoader("hfhub_errors", globals(), "huggingface_hub.errors")
3637

3738
_CONFIG_REGISTRY_OVERRIDE_HF: Dict[str, str] = {
3839
"mllama": "MllamaConfig"
3940
}
4041

41-
_CONFIG_REGISTRY: Dict[str, Type[PretrainedConfig]] = {
42-
"chatglm": ChatGLMConfig,
43-
"cohere2": Cohere2Config,
44-
"dbrx": DbrxConfig,
45-
"deepseek_vl_v2": DeepseekVLV2Config,
46-
"kimi_vl": KimiVLConfig,
47-
"mpt": MPTConfig,
48-
"RefinedWeb": RWConfig, # For tiiuae/falcon-40b(-instruct)
49-
"RefinedWebModel": RWConfig, # For tiiuae/falcon-7b(-instruct)
50-
"jais": JAISConfig,
51-
"mlp_speculator": MLPSpeculatorConfig,
52-
"medusa": MedusaConfig,
53-
"eagle": EAGLEConfig,
54-
"exaone": ExaoneConfig,
55-
"h2ovl_chat": H2OVLChatConfig,
56-
"internvl_chat": InternVLChatConfig,
57-
"nemotron": NemotronConfig,
58-
"NVLM_D": NVLM_D_Config,
59-
"olmo2": Olmo2Config,
60-
"solar": SolarConfig,
61-
"skywork_chat": SkyworkR1VChatConfig,
62-
"telechat": Telechat2Config,
63-
"ultravox": UltravoxConfig,
42+
_CONFIG_REGISTRY: Dict[str, str] = {
43+
"chatglm": "ChatGLMConfig",
44+
"cohere2": "Cohere2Config",
45+
"dbrx": "DbrxConfig",
46+
"deepseek_vl_v2": "DeepseekVLV2Config",
47+
"kimi_vl": "KimiVLConfig",
48+
"mpt": "MPTConfig",
49+
"RefinedWeb": "RWConfig", # For tiiuae/falcon-40b(-instruct)
50+
"RefinedWebModel": "RWConfig", # For tiiuae/falcon-7b(-instruct)
51+
"jais": "JAISConfig",
52+
"mlp_speculator": "MLPSpeculatorConfig",
53+
"medusa": "MedusaConfig",
54+
"eagle": "EAGLEConfig",
55+
"exaone": "ExaoneConfig",
56+
"h2ovl_chat": "H2OVLChatConfig",
57+
"internvl_chat": "InternVLChatConfig",
58+
"nemotron": "NemotronConfig",
59+
"NVLM_D": "NVLM_D_Config",
60+
"olmo2": "Olmo2Config",
61+
"solar": "SolarConfig",
62+
"skywork_chat": "SkyworkR1VChatConfig",
63+
"telechat": "Telechat2Config",
64+
"ultravox": "UltravoxConfig",
6465
**_CONFIG_REGISTRY_OVERRIDE_HF
6566
}
6667

@@ -371,7 +372,7 @@ def try_get_local_file(model: Union[str, Path],
371372
revision=revision)
372373
if isinstance(cached_filepath, str):
373374
return Path(cached_filepath)
374-
except hfhub_utils.HFValidationError:
375+
except hfhub_errors.HFValidationError:
375376
...
376377
return None
377378

0 commit comments

Comments
 (0)