Skip to content

DeepSeek-R1-Distill-Qwen-32B-GGUF needs the deepseek-r1-qwen tokenizerΒ #1900

Closed
@Kenshiro-28

Description

@Kenshiro-28

Current Behavior

Can't run DeepSeek-R1-Distill-Qwen-32B-GGUF, I get a tokenizer error:

"llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'deepseek-r1-qwen'"

Model url: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF

Environment and Context

Hardware: VPS with 32 GB RAM
OS: Debian 12

Using the current version of llama-cpp-python

Suggestion

Upgrade to the required version of llama.cpp:

https://github.com/ggerganov/llama.cpp/releases/tag/b4514

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions