-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support VLM model and GPT4V API #2058
Comments
I strongly second this. However Llava 1.6 models are very peculiar :D |
It's fine if it's peculiar. We can just error (BadRequest) anything that doesn't have image first or multiple images for instance, but still keep the same format. We shouldn't design the endpoint around the model. The model should fit into the endpoint. |
7 tasks
3 tasks
Closing this as we merged #5237 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
VLM model: Qwen-VL/LLaVA etc
VLM API: GPT-4V API https://platform.openai.com/docs/guides/vision
The text was updated successfully, but these errors were encountered: