-
Notifications
You must be signed in to change notification settings - Fork 964
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add reranking support #1794
Comments
@abetlen Sorry but do you have any plans to implement this? |
Hi @donguyen32 |
@yutyan0119 Arcoding from the original repo, I see that the format of the rerank task is |
@donguyen32 I have verified that the output is the same as the original implementation with the following command ./llama-embedding \
-m models/bge-reranker-v2-m3/ggml-model-f16.gguf \
-p "what is panda?</s><s>hi\nwhat is panda?</s><s>it's a bear\nwhat is panda?</s><s>The giant panda (Ailuropoda melanoleuca), sometimes called a panda bear or simply panda, is a bear species endemic to China." \
--pooling rank --embd-normalize -1 --verbose-prompt The same command also seems to be used for testing on CI. In fact, I do not know how these symbols affect the accuracy of Rerank. If you know, please let me know. And if we want a return value in the form of the server, I think it would be better to have a separate method in the form of a create_embedding method for embed, like |
According to the ggerganov/llama.cpp#9510, lllama-cpp supported for reranking model https://huggingface.co/BAAI/bge-reranker-v2-m3.
Please provide support for this version.
The text was updated successfully, but these errors were encountered: