Skip to content

Commit c08b543

Browse files
csy1204xuebwang-amd
authored andcommitted
[CI/Build] upgrade compressed-tensors to 0.12.2 to address LGPLv3 (vllm-project#26501)
Signed-off-by: Sangyeon Cho <josang1204@gmail.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
1 parent b99be4b commit c08b543

File tree

3 files changed

+3
-3
lines changed

3 files changed

+3
-3
lines changed

requirements/common.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ pyyaml
3838
six>=1.16.0; python_version > '3.11' # transitive dependency of pandas that needs to be the latest version for python 3.12
3939
setuptools>=77.0.3,<80; python_version > '3.11' # Setuptools is used by triton, we need to ensure a modern version is installed for 3.12+ so that it does not try to import distutils, which was removed in 3.12
4040
einops # Required for Qwen2-VL.
41-
compressed-tensors == 0.11.0 # required for compressed-tensors
41+
compressed-tensors == 0.12.2 # required for compressed-tensors
4242
depyf==0.19.0 # required for profiling and debugging with compilation config
4343
cloudpickle # allows pickling lambda functions in model_executor/models/registry.py
4444
watchfiles # required for http server to monitor the updates of TLS files

vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -310,7 +310,7 @@ def _is_fp4a4_nvfp4(
310310
)
311311
is_float_type = (
312312
weight_quant.type == QuantizationType.FLOAT
313-
and input_quant.type == QuantizationType.FLOAT.value
313+
and input_quant.type == QuantizationType.FLOAT
314314
)
315315
is_4_bits = weight_quant.num_bits == 4 and input_quant.num_bits == 4
316316

vllm/model_executor/layers/quantization/compressed_tensors/compressed_tensors_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -143,7 +143,7 @@ def get_moe_method(
143143
# Prefer to use the MarlinMoE kernel when it is supported.
144144
if not check_moe_marlin_supports_layer(layer, group_size):
145145
if (
146-
weight_quant.strategy in QuantizationStrategy.GROUP
146+
weight_quant.strategy == QuantizationStrategy.GROUP
147147
and weight_quant.actorder
148148
in (ActivationOrdering.GROUP, ActivationOrdering.DYNAMIC)
149149
):

0 commit comments

Comments
 (0)