File tree 1 file changed +9
-2
lines changed
1 file changed +9
-2
lines changed Original file line number Diff line number Diff line change @@ -1037,9 +1037,17 @@ def load_model(self) -> None:
1037
1037
assert supports_lora (
1038
1038
self .model
1039
1039
), f"{ self .model .__class__ .__name__ } does not support LoRA yet."
1040
+
1040
1041
if supports_multimodal (self .model ):
1041
1042
logger .warning ("Regarding multimodal models, vLLM currently "
1042
1043
"only supports adding LoRA to language model." )
1044
+ # It's necessary to distinguish between the max_position_embeddings
1045
+ # of VLMs and LLMs.
1046
+ if hasattr (self .model .config , "max_position_embeddings" ):
1047
+ max_pos_embeddings = self .model .config .max_position_embeddings
1048
+ else :
1049
+ max_pos_embeddings = (
1050
+ self .model .config .text_config .max_position_embeddings )
1043
1051
1044
1052
self .lora_manager = LRUCacheWorkerLoRAManager (
1045
1053
self .scheduler_config .max_num_seqs ,
@@ -1049,8 +1057,7 @@ def load_model(self) -> None:
1049
1057
self .device ,
1050
1058
self .model .embedding_modules ,
1051
1059
self .model .embedding_padding_modules ,
1052
- max_position_embeddings = self .model .config .
1053
- max_position_embeddings ,
1060
+ max_position_embeddings = max_pos_embeddings ,
1054
1061
)
1055
1062
self .model = self .lora_manager .create_lora_manager (self .model )
1056
1063
You can’t perform that action at this time.
0 commit comments