Skip to content

[FEATURE] Support batch inference  #1840

@ylwu-amzn

Description

@ylwu-amzn

Most model services have some throttling limit. For example Bedrock.

With such limit, it takes long time to ingest large amount of data. One way is to use batch inference to increase the throughput. For example, Bedrock supports batch inference https://docs.aws.amazon.com/bedrock/latest/userguide/batch-inference.html

Metadata

Metadata

Assignees

Labels

enhancementNew feature or request

Type

No type

Projects

Status

Done

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions