File tree Expand file tree Collapse file tree 1 file changed +9
-2
lines changed Expand file tree Collapse file tree 1 file changed +9
-2
lines changed Original file line number Diff line number Diff line change @@ -1069,9 +1069,17 @@ def load_model(self) -> None:
1069
1069
assert supports_lora (
1070
1070
self .model
1071
1071
), f"{ self .model .__class__ .__name__ } does not support LoRA yet."
1072
+
1072
1073
if supports_multimodal (self .model ):
1073
1074
logger .warning ("Regarding multimodal models, vLLM currently "
1074
1075
"only supports adding LoRA to language model." )
1076
+ # It's necessary to distinguish between the max_position_embeddings
1077
+ # of VLMs and LLMs.
1078
+ if hasattr (self .model .config , "max_position_embeddings" ):
1079
+ max_pos_embeddings = self .model .config .max_position_embeddings
1080
+ else :
1081
+ max_pos_embeddings = (
1082
+ self .model .config .text_config .max_position_embeddings )
1075
1083
1076
1084
self .lora_manager = LRUCacheWorkerLoRAManager (
1077
1085
self .scheduler_config .max_num_seqs ,
@@ -1081,8 +1089,7 @@ def load_model(self) -> None:
1081
1089
self .device ,
1082
1090
self .model .embedding_modules ,
1083
1091
self .model .embedding_padding_modules ,
1084
- max_position_embeddings = self .model .config .
1085
- max_position_embeddings ,
1092
+ max_position_embeddings = max_pos_embeddings ,
1086
1093
)
1087
1094
self .model = self .lora_manager .create_lora_manager (self .model )
1088
1095
You can’t perform that action at this time.
0 commit comments