Skip to content

Commit

Permalink
expose the cabability of choosing DL framework of the HF pipelines mo… (
Browse files Browse the repository at this point in the history
  • Loading branch information
saeid93 authored Apr 4, 2023
1 parent 84cbe8b commit 7d8a207
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions runtimes/huggingface/mlserver_huggingface/common.py
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,7 @@ class Config:
task_suffix: str = ""
pretrained_model: Optional[str] = None
pretrained_tokenizer: Optional[str] = None
framework: Optional[str] = None
optimum_model: bool = False
device: int = -1

Expand Down Expand Up @@ -140,6 +141,7 @@ def load_pipeline_from_settings(
tokenizer=tokenizer,
device=device,
batch_size=batch_size,
framework=hf_settings.framework,
)

# If max_batch_size > 0 we need to ensure tokens are padded
Expand Down

0 comments on commit 7d8a207

Please sign in to comment.