File tree 1 file changed +5
-5
lines changed 1 file changed +5
-5
lines changed Original file line number Diff line number Diff line change @@ -39,15 +39,15 @@ python ./examples/llava/llava_surgery.py -m path/to/MobileVLM-1.7B
39
39
3 . Use ` convert_image_encoder_to_gguf.py ` with ` --projector-type ldp ` (for ** V2** please use ` --projector-type ldpv2 ` ) to convert the LLaVA image encoder to GGUF:
40
40
41
41
``` sh
42
- python ./examples/llava/convert_image_encoder_to_gguf \
42
+ python ./examples/llava/convert_image_encoder_to_gguf.py \
43
43
-m path/to/clip-vit-large-patch14-336 \
44
44
--llava-projector path/to/MobileVLM-1.7B/llava.projector \
45
45
--output-dir path/to/MobileVLM-1.7B \
46
46
--projector-type ldp
47
47
```
48
48
49
49
``` sh
50
- python ./examples/llava/convert_image_encoder_to_gguf \
50
+ python ./examples/llava/convert_image_encoder_to_gguf.py \
51
51
-m path/to/clip-vit-large-patch14-336 \
52
52
--llava-projector path/to/MobileVLM-1.7B_V2/llava.projector \
53
53
--output-dir path/to/MobileVLM-1.7B_V2 \
@@ -57,12 +57,12 @@ python ./examples/llava/convert_image_encoder_to_gguf \
57
57
4 . Use ` examples/convert_legacy_llama.py ` to convert the LLaMA part of LLaVA to GGUF:
58
58
59
59
``` sh
60
- python ./examples/convert_legacy_llama.py path/to/MobileVLM-1.7B
60
+ python ./examples/convert_legacy_llama.py path/to/MobileVLM-1.7B --skip-unknown
61
61
```
62
62
63
- 5 . Use ` quantize ` to convert LLaMA part's DataType from ` fp16 ` to ` q4_k `
63
+ 5 . Use ` quantize ` to convert LLaMA part's DataType from ` fp32 ` to ` q4_k `
64
64
``` sh
65
- ./llama-quantize path/to/MobileVLM-1.7B/ggml-model-f16 .gguf path/to/MobileVLM-1.7B/ggml-model-q4_k.gguf q4_k_s
65
+ ./llama-quantize path/to/MobileVLM-1.7B/ggml-model-F32 .gguf path/to/MobileVLM-1.7B/ggml-model-q4_k.gguf q4_k_s
66
66
```
67
67
68
68
Now both the LLaMA part and the image encoder is in the ` MobileVLM-1.7B ` directory.
You can’t perform that action at this time.
0 commit comments