-
Notifications
You must be signed in to change notification settings - Fork 29.9k
[aya vision] fix processor for vLLM #38371
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This behavior seems quite inconsistent across models. I found that most processors are able to take extra For the processors that perform this check more strictly, in vLLM we override Example: https://github.com/vllm-project/vllm/blob/main/vllm/model_executor/models/idefics3.py#L321-L325 |
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
Ah oke, that makes sense. We haven't been consistent in checking placeholder tokens in processors, and each model contributor added the model as it was in original implementation I think in that case I can't add a robust test. I believe integration with transformers backend will auto-resolve future issues, I am planning to get it ready in this week |
Sounds good, thanks for fixing! |
Thanks for adding the |
Looks like this PR didn't make it into the release :( |
Hmm, cc @ArthurZucker. On main branch it is fine, since I merged last Friday the last PR for vLLM. The last patch isn't though 🥲 |
It did but the diff was empty |
What does this PR do?
Fixes #38350
@DarkLight1337 , one quick question. I am trying to add a test on our side, so we don't break anything that vLLM relies on. For the case of aya-vision the error log shows inputs as (
text="<image>, images=None"
) but doing so will fail on other processors where we have extra checks that "num input images == num image tokens
"I wonder how that is bypassed on vLLM and how can I add a test to cover assumptions vLLM has about our processors