Skip to content

Commit ee3faa9

Browse files
authored
Merge pull request vllm-project#12 from dcmaddix/fused_moe_lora_test
Add support for FSDP lora adapter format from PEFT
2 parents 0687656 + 9935039 commit ee3faa9

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

vllm/lora/models.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -216,6 +216,10 @@ def check_unexpected_modules(modules: dict):
216216
for lora_module in modules.keys(): # noqa
217217
module_name, _ = parse_fine_tuned_lora_name(lora_module, weights_mapper)
218218
part_name = module_name.split(".")[-1]
219+
# Handle FSDP file format where experts.base_layer is the
220+
# gate_up_proj and experts is the down_proj
221+
if "base_layer" in lora_module:
222+
continue
219223
if part_name not in expected_lora_modules:
220224
unexpected_modules.append(module_name)
221225
if unexpected_modules:

0 commit comments

Comments
 (0)