-
-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Introduce return_prompt
request option to api_server
entrypoint
#1232
Introduce return_prompt
request option to api_server
entrypoint
#1232
Conversation
return_prompt
option to api_server
entrypointreturn_prompt
request option to api_server
entrypoint
Please consider implementing the See: https://platform.openai.com/docs/api-reference/completions/create |
Add simple test cases to cover the flags. |
I will add it. |
I'm considering it. Should I add it in the same PR or new one? |
Same PR. They belong together, same feature. |
Thank you for your contribution! Our goal of |
@zhuohan123 The addition does not look complex at all. At least not complex as Please let me know if I can open a new PR. Thanks anyway! |
这个问题解决了吗?怎么刚测试还有了? |
can use res["text"][0].replace(promopt,"") to solve this problem?? |
"can use res["text"][0].replace(promopt,"") to solve this problem??" You can in simple examples, but for a RAG workflow this is not going to work as you can be sending complex data that will be formatted accordingly when sent and therefore won't be the same when it is returned. The only way to do that would be to change every document in your RAG workflow to mirror the prompt. I see this as closed, but I don't see "return_prompt" in the uses for the vllm api server?
|
Hello @Kaotic3! This issue is closed because the vllm/vllm/entrypoints/api_server.py Lines 2 to 6 in 429284d
|
Does this mean that we don't get the echo command then? It isn't really even a problem with vLLM it is a problem with outlines served via vLLM - but I figured the flags are the same on both, so if it existed here, then I could use it in outlines. |
The |
It isn't as that list is from the openai compatible server. --echo isn't in the list. Just to add because I know it might be confusing, the command for the openai compatible is:
So it is still api_server.py - even when using the openai version. |
You don't pass it as an argument when launching the server. It should be specified in your payload to the vllm/vllm/entrypoints/openai/serving_completion.py Lines 231 to 242 in 429284d
|
I see, it is false by default so not really utilised it before. I will try it with outlines, thanks man appreciate the assistance! |
Introduce
return_prompt
request option toapi_server
entrypoint. The default value forreturn_prompt
isFalse
.Issue: #1043