feat(model): Support llama.cpp server deploy #2263
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
/v1/completions
How Has This Been Tested?
Install dependencies
pip install -e ".[llama_cpp_server]"
If you want to accelerate the inference speed, and you have a GPU, you can install the following dependencies:
Download the model
Here, we use the
qwen2.5-0.5b-instruct
model as an example. You can download the model from the Huggingface.wget https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/qwen2.5-0.5b-instruct-q4_k_m.gguf?download=true -O /tmp/qwen2.5-0.5b-instruct-q4_k_m.gguf
Modify configuration file
In the
.env
configuration file, modify the inference type of the model to startllama.cpp
inference.Start the DB-GPT server
Snapshots:
Include snapshots for easier review.
Checklist: