-
Notifications
You must be signed in to change notification settings - Fork 37
[model] support vision language model llava. #178
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
bfa4b3c
to
45db749
Compare
1fb4495
to
68013b4
Compare
96fb59b
to
0ea99b6
Compare
7182aad
to
4e99594
Compare
3c41efc
to
6dd9cea
Compare
scalellm/_C/vlm_handler.pyi
Outdated
from scalellm._C.output import RequestOutput | ||
from scalellm._C.sampling_params import SamplingParams | ||
|
||
class Priority(Enum): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
redefinition error for Priority.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
fixed
@@ -49,6 +49,8 @@ void Batch::add(Sequence* sequence, uint32_t token_budget) { | |||
sequences_.push_back(sequence); | |||
token_budgets_.push_back(token_budget); | |||
budget_used_.push_back(0); | |||
|
|||
input_embedding_ = sequence->get_input_embedding(); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
sounds one input embedding for each sequence, if so, it should be a vector of input embeddings for the batch.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yes, in this pr only support one sequence + one image, and will support multiple sequences later.
LGTM. Thank you for your hard work in getting the first VLM up and running. I've left some comments; I believe we should refactor to reduce redundant code between the worker and engine after landing this. |
3daa679
to
137360a
Compare
137360a
to
baa26a4
Compare
Let's revert the VLM related change for now to unblock release. cc: @liutongxuan - [ ] @guocuimi to create a branch for VLM development. This reverts commit 437be3f and d711c55.
(cherry picked from commit 437be3f)
No description provided.