Skip to content

[model] support vision language model llava. #178

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 9 commits into from
Jun 28, 2024

Conversation

liutongxuan
Copy link
Contributor

No description provided.

@liutongxuan liutongxuan force-pushed the features/llava branch 3 times, most recently from 1fb4495 to 68013b4 Compare May 28, 2024 05:08
@liutongxuan liutongxuan changed the title [feat] support vision language model llava. [model] support vision language model llava. Jun 4, 2024
@liutongxuan liutongxuan force-pushed the features/llava branch 10 times, most recently from 7182aad to 4e99594 Compare June 26, 2024 14:25
@liutongxuan liutongxuan force-pushed the features/llava branch 6 times, most recently from 3c41efc to 6dd9cea Compare June 27, 2024 09:54
from scalellm._C.output import RequestOutput
from scalellm._C.sampling_params import SamplingParams

class Priority(Enum):
Copy link
Collaborator

@guocuimi guocuimi Jun 28, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

redefinition error for Priority.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

fixed

@@ -49,6 +49,8 @@ void Batch::add(Sequence* sequence, uint32_t token_budget) {
sequences_.push_back(sequence);
token_budgets_.push_back(token_budget);
budget_used_.push_back(0);

input_embedding_ = sequence->get_input_embedding();
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

sounds one input embedding for each sequence, if so, it should be a vector of input embeddings for the batch.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yes, in this pr only support one sequence + one image, and will support multiple sequences later.

@guocuimi
Copy link
Collaborator

LGTM. Thank you for your hard work in getting the first VLM up and running. I've left some comments; I believe we should refactor to reduce redundant code between the worker and engine after landing this.

@liutongxuan liutongxuan force-pushed the features/llava branch 4 times, most recently from 3daa679 to 137360a Compare June 28, 2024 06:20
@liutongxuan liutongxuan merged commit 437be3f into vectorch-ai:main Jun 28, 2024
4 checks passed
guocuimi added a commit that referenced this pull request Jul 3, 2024
guocuimi added a commit that referenced this pull request Jul 3, 2024
guocuimi added a commit that referenced this pull request Jul 3, 2024
Let's revert the VLM related change for now to unblock release. cc:
@liutongxuan

- [ ] @guocuimi to create a branch for VLM development.

This reverts commit 437be3f and
d711c55.
guocuimi pushed a commit that referenced this pull request Jul 3, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants