We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 5a60699 commit 1bf2dd9Copy full SHA for 1bf2dd9
vllm/model_executor/models/gemma2.py
@@ -312,6 +312,14 @@ class Gemma2ForCausalLM(nn.Module, SupportsLoRA):
312
# Gemma does not apply LoRA to the embedding layer.
313
embedding_modules = {}
314
embedding_padding_modules = []
315
+ bitsandbytes_stacked_params_mapping = {
316
+ # shard_name, weight_name, index
317
+ "q_proj": ("qkv_proj", 0),
318
+ "k_proj": ("qkv_proj", 1),
319
+ "v_proj": ("qkv_proj", 2),
320
+ "gate_proj": ("gate_up_proj", 0),
321
+ "up_proj": ("gate_up_proj", 1),
322
+ }
323
324
def __init__(
325
self,
0 commit comments