Closed
Description
When I use the ./quantize command to use the Q4_K_M parameter and want to convert a vector file, an error is reported
main: quantizing './zh-models/plus_13B/ggml-model-f16.bin' to './zh-models/plus_13B/ggml-model-q4_K_M.bin' as Q4_K_M
llama.cpp: loading model from ./zh-models/plus_13B/ggml-model-f16.bin
llama.cpp: saving model to ./zh-models/plus_13B/ggml-model-q4_K_M.bin
========================= Tensor sizes 5120 x 49954 are not divisible by 256
This is required to be able to use k-quants for now!
========================================================================================
llama_model_quantize: failed to quantize: Unsupported tensor size encountered
Can you tell me what might be wrong?
env
max os ventura 13.4.1
Metadata
Metadata
Assignees
Labels
No labels