Skip to content

Commit

Permalink
fix typo; add test
Browse files Browse the repository at this point in the history
  • Loading branch information
dsikka committed Sep 30, 2024
1 parent 3d12554 commit b54b633
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 1 deletion.
1 change: 1 addition & 0 deletions tests/weight_loading/models-large.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,3 +2,4 @@ compressed-tensors, nm-testing/Mixtral-8x7B-Instruct-v0.1-W4A16-quantized, main
compressed-tensors, nm-testing/Mixtral-8x7B-Instruct-v0.1-W4A16-channel-quantized, main
compressed-tensors, nm-testing/Mixtral-8x7B-Instruct-v0.1-W8A16-quantized, main
gptq_marlin, TheBloke/Mixtral-8x7B-v0.1-GPTQ, main
awq_marlin, casperhansen/deepseek-coder-v2-instruct-awq, main
2 changes: 1 addition & 1 deletion vllm/model_executor/layers/quantization/awq_marlin.py
Original file line number Diff line number Diff line change
Expand Up @@ -293,7 +293,7 @@ def create_weights(self, layer: torch.nn.Module, num_experts: int,
"is_transposed":
True,
"quant_method":
FusedMoeWeightScaleSupported.GROUP,
FusedMoeWeightScaleSupported.GROUP.value,
})

w13_qweight = Parameter(torch.empty(num_experts,
Expand Down

0 comments on commit b54b633

Please sign in to comment.