Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bugfix] Fix LoRA loading check #4138

Merged
merged 8 commits into from
Apr 19, 2024

Conversation

jeejeelee
Copy link
Contributor

Following #3777. Considering that peft can specify the target_modules as follows, we need to optimize the LoRa loading check.

{
  "alpha_pattern": {},
  "auto_mapping": null,
  "base_model_name_or_path": "Llama-2-7B-fp16-hf",
  "bias": "none",
  "fan_in_fan_out": false,
  "inference_mode": true,
  "init_lora_weights": true,
  "layers_pattern": null,
  "layers_to_transform": null,
  "loftq_config": {},
  "lora_alpha": 32,
  "lora_dropout": 0.0,
  "megatron_config": null,
  "megatron_core": "megatron.core",
  "modules_to_save": null,
  "peft_type": "LORA",
  "r": 8,
  "rank_pattern": {},
  "revision": null,
  "target_modules": [
    "model.layers.11.self_attn.k_proj",
    "model.layers.0.self_attn.q_proj",
    "model.layers.2.self_attn.o_proj"
  ],
  "task_type": "CAUSAL_LM",
  "use_dora": false,
  "use_rslora": false
}

Copy link
Collaborator

@Yard1 Yard1 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can we add a unit test?

@jeejeelee
Copy link
Contributor Author

@Yard1 Thanks for review, I have added a unit test for this

@jeejeelee jeejeelee requested a review from Yard1 April 18, 2024 11:50
@jeejeelee jeejeelee mentioned this pull request Apr 19, 2024
9 tasks
@simon-mo simon-mo enabled auto-merge (squash) April 19, 2024 06:01
@simon-mo simon-mo disabled auto-merge April 19, 2024 07:59
@simon-mo simon-mo merged commit d17c847 into vllm-project:main Apr 19, 2024
44 of 47 checks passed
robertgshaw2-neuralmagic pushed a commit to neuralmagic/nm-vllm that referenced this pull request Apr 21, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
xjpang pushed a commit to xjpang/vllm that referenced this pull request Apr 25, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
xjpang pushed a commit to xjpang/vllm that referenced this pull request Apr 26, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
robertgshaw2-neuralmagic pushed a commit to neuralmagic/nm-vllm that referenced this pull request Apr 26, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
alexeykondrat pushed a commit to alexeykondrat/ci-vllm that referenced this pull request May 1, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
z103cb pushed a commit to z103cb/opendatahub_vllm that referenced this pull request May 7, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
@jeejeelee jeejeelee deleted the optimize-lora-load branch August 19, 2024 08:09
Temirulan pushed a commit to Temirulan/vllm-whisper that referenced this pull request Sep 6, 2024
Co-authored-by: simon-mo <simon.mo@hey.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants