Skip to content

Conversation

@ashwin-phadke
Copy link
Contributor

@ashwin-phadke ashwin-phadke commented Sep 27, 2025

#23892 : Removes support for LoRA bias completely from vLLM. LoRA bias was added in #5733.

Looks like layers.py has been kind of replaced with the layers directory, this needs some review.

FIX #23892

Purpose

Test Plan

  1. Update all tests that reference LoRA bias and re run them

Test Result


Essential Elements of an Effective PR Description Checklist

In progress:

  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
@mergify mergify bot added the tpu Related to Google TPUs label Sep 27, 2025
@ashwin-phadke ashwin-phadke mentioned this pull request Sep 27, 2025
3 tasks
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
@@ -60,17 +60,97 @@

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Don't modify unrelated code

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This seems like a bad merge which entirely reverts all the configu modularisation work

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, I believe it is a bad merge too, working on it.

@mergify
Copy link

mergify bot commented Sep 30, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @ashwin-phadke.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase label Sep 30, 2025
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
@mergify mergify bot removed the needs-rebase label Sep 30, 2025
@ashwin-phadke
Copy link
Contributor Author

This should solve it.

@hmellor hmellor requested a review from jeejeelee September 30, 2025 17:05
@ashwin-phadke
Copy link
Contributor Author

@jeejeelee @hmellor any feedback on this one?
I noticed my comments on section x removed made it through, so I will remove those, other than that do you guys see any issues that must be addressed? anything missing or done incorrectly?
Let me know.

@jeejeelee
Copy link
Collaborator

Very sorry for the delayed feedback due to the holiday. Can you resolve these conflicts?

@hmellor
Copy link
Member

hmellor commented Oct 7, 2025

See https://vllm-dev.slack.com/archives/C07R5Q1Q2BB/p1759663228844749 for detailed instructions on resolving the conflicts

@mergify
Copy link

mergify bot commented Oct 9, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @ashwin-phadke.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase label Oct 9, 2025
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Copy link
Collaborator

@jeejeelee jeejeelee left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ashwin-phadke Thank you for contribution.

@jeejeelee
Copy link
Collaborator

@Yikun @xuechendi We are cleaning up lora bias related code (see: #23892), which may lead to the plugin LoRA failure

@jeejeelee jeejeelee added the ready ONLY add when PR is ready to merge/full CI is needed label Oct 10, 2025
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
@jeejeelee jeejeelee enabled auto-merge (squash) October 10, 2025 09:20
@jeejeelee jeejeelee merged commit ab196ed into vllm-project:main Oct 10, 2025
55 checks passed
Dhruvilbhatt pushed a commit to Dhruvilbhatt/vllm that referenced this pull request Oct 14, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Dhruvil Bhatt <bhattdbh@amazon.com>
bbartels pushed a commit to bbartels/vllm that referenced this pull request Oct 16, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: bbartels <benjamin@bartels.dev>
MengqingCao added a commit to MengqingCao/vllm-ascend-fork that referenced this pull request Oct 17, 2025
Signed-off-by: MengqingCao <cmq0113@163.com>
MengqingCao pushed a commit to MengqingCao/vllm-ascend-fork that referenced this pull request Oct 18, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
MengqingCao pushed a commit to MengqingCao/vllm-ascend-fork that referenced this pull request Oct 19, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
MengqingCao pushed a commit to MengqingCao/vllm-ascend-fork that referenced this pull request Oct 20, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
wxsIcey added a commit to wxsIcey/vllm-ascend that referenced this pull request Oct 20, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
wxsIcey added a commit to wxsIcey/vllm-ascend that referenced this pull request Oct 20, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
Signed-off-by: Icey <1790571317@qq.com>
wxsIcey added a commit to wxsIcey/vllm-ascend that referenced this pull request Oct 20, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
Signed-off-by: Icey <1790571317@qq.com>
lywa1998 pushed a commit to lywa1998/vllm that referenced this pull request Oct 20, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
MengqingCao pushed a commit to MengqingCao/vllm-ascend-fork that referenced this pull request Oct 20, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
MengqingCao pushed a commit to MengqingCao/vllm-ascend-fork that referenced this pull request Oct 20, 2025
  * fix bert model
  * fix guided decoding
  * revert skipped e2e test
  * fix lora vllm-project/vllm#25807
  * fix vl

Signed-off-by: MengqingCao <cmq0113@163.com>
alhridoy pushed a commit to alhridoy/vllm that referenced this pull request Oct 24, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
wangxiyuan pushed a commit to vllm-project/vllm-ascend that referenced this pull request Oct 24, 2025
### What this PR does / why we need it?
This is the step 1 of refactoring code to adapt with vllm main, and this
pr aligned with
vllm-project/vllm@17c540a

1. refactor deepseek to the latest code arch as of
vllm-project/vllm@17c540a
 
2. bunches of fixes due to vllm changes
- Fix `AscendScheduler` `__post_init__`, caused by
vllm-project/vllm#25075
- Fix `AscendScheduler` init got an unexpected arg `block_size`, caused
by vllm-project/vllm#26296
- Fix `KVCacheManager` `get_num_common_prefix_blocks` arg, caused by
vllm-project/vllm#23485
- Fix `MLAAttention` import,caused by
vllm-project/vllm#25103
- Fix `SharedFusedMoE` import, caused by
vllm-project/vllm#26145
- Fix `LazyLoader` improt, caused by
vllm-project/vllm#27022
- Fix `vllm.utils.swap_dict_values` improt, caused by
vllm-project/vllm#26990
- Fix `Backend` enum import, caused by
vllm-project/vllm#25893
- Fix `CompilationLevel` renaming to `CompilationMode` issue introduced
by vllm-project/vllm#26355
- Fix fused_moe ops, caused by
vllm-project/vllm#24097
- Fix bert model because of `inputs_embeds`, caused by
vllm-project/vllm#25922
- Fix MRope because of `get_input_positions_tensor` to
`get_mrope_input_positions`, caused by
vllm-project/vllm#24172
- Fix `splitting_ops` changes introduced by
vllm-project/vllm#25845
- Fix multi-modality changes introduced by
vllm-project/vllm#16229
- Fix lora bias dropping issue introduced by
vllm-project/vllm#25807
- Fix structured ouput break introduced by
vllm-project/vllm#26737

### Does this PR introduce _any_ user-facing change?

### How was this patch tested?
CI passed with existing test.


- vLLM version: v0.11.0rc3
- vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0

---------

Signed-off-by: MengqingCao <cmq0113@163.com>
Signed-off-by: Icey <1790571317@qq.com>
Co-authored-by: Icey <1790571317@qq.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 24, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
0xrushi pushed a commit to 0xrushi/vllm that referenced this pull request Oct 26, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
0xrushi pushed a commit to 0xrushi/vllm that referenced this pull request Oct 26, 2025
Signed-off-by: Ashwin Phadke <ashwinphadke12@rediffmail.com>
Signed-off-by: Ashwin Phadke <23502062+ashwin-phadke@users.noreply.github.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed tpu Related to Google TPUs

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[RFC]: Remove LoRA bias

3 participants