-
Notifications
You must be signed in to change notification settings - Fork 102
feat: add mm embed worker and service. #518
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
feat: add mm embed worker and service. #518
Conversation
| // if raw_output.outputs.size() value is 0, | ||
| // this means all sequences are in prefill stage status. | ||
| const int64_t num_seqs = raw_output.outputs.size(); | ||
| int64_t num_seqs; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
还是需要加上MMBatch的子类,不能改变现有Batch逻辑。
d642170 to
a103d80
Compare
fix: convert bytes data to base64 encoding. fix: fix base64 encoding.
| MMBatchData(const std::vector<MMData>& datas); | ||
| MMBatchData(uint32_t ty, const MMDict& items); | ||
|
|
||
| bool has(uint32_t type) const { return type & ty_ != 0; } |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
type & (ty_ != 0) may be better.
| if (model_backend == "llm") { | ||
| worker_type = | ||
| (options.task_type() == "generate") ? WorkerType::LLM : WorkerType::ELM; | ||
| if (options.task_type() == "gnerate") { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
generate may be ok?
…6 and bf16 types.
add mm embedding service.
add mm embedding worker.
add tensor serialization and deserialization in shared memory manager.