Skip to content

Commit

Permalink
Add default for max_seq_len
Browse files Browse the repository at this point in the history
  • Loading branch information
ashwinb committed Sep 19, 2024
1 parent 9ab27e8 commit f5eda1d
Showing 1 changed file with 3 additions and 3 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -11,10 +11,10 @@
from llama_models.schema_utils import json_schema_type
from llama_models.sku_list import all_registered_models, resolve_model

from llama_stack.apis.inference import QuantizationConfig

from pydantic import BaseModel, Field, field_validator

from llama_stack.apis.inference import QuantizationConfig


@json_schema_type
class MetaReferenceImplConfig(BaseModel):
Expand All @@ -24,7 +24,7 @@ class MetaReferenceImplConfig(BaseModel):
)
quantization: Optional[QuantizationConfig] = None
torch_seed: Optional[int] = None
max_seq_len: int
max_seq_len: int = 4096
max_batch_size: int = 1

@field_validator("model")
Expand Down

0 comments on commit f5eda1d

Please sign in to comment.