File tree Expand file tree Collapse file tree 2 files changed +5
-1
lines changed
model_executor/layers/fused_moe Expand file tree Collapse file tree 2 files changed +5
-1
lines changed Original file line number Diff line number Diff line change 41
41
from .pplx_prepare_finalize import PplxPrepareAndFinalize
42
42
else :
43
43
fused_experts = None # type: ignore
44
+ FusedMoEPermuteExpertsUnpermute = None # type: ignore
44
45
FusedMoEPrepareAndFinalize = None # type: ignore
45
46
if is_rocm_aiter_moe_enabled ():
46
47
from vllm .model_executor .layers .fused_moe .rocm_aiter_fused_moe import ( # noqa: E501
Original file line number Diff line number Diff line change @@ -300,7 +300,10 @@ def get_config(
300
300
" - For Hugging Face models: ensure the presence of a "
301
301
"'config.json'.\n "
302
302
" - For Mistral models: ensure the presence of a "
303
- "'params.json'.\n " ).format (model = model )
303
+ "'params.json'.\n "
304
+ "3. For GGUF: pass the local path of the GGUF checkpoint.\n "
305
+ " Loading GGUF from a remote repo directly is not yet "
306
+ "supported.\n " ).format (model = model )
304
307
305
308
raise ValueError (error_message ) from e
306
309
You can’t perform that action at this time.
0 commit comments