We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 1d0bdfa commit 70fbee3Copy full SHA for 70fbee3
OAI/types/model.py
@@ -6,7 +6,7 @@
6
class ModelCardParameters(BaseModel):
7
max_seq_len: Optional[int] = 4096
8
rope_scale: Optional[float] = 1.0
9
- rope_alpha: Optional[float] = None
+ rope_alpha: Optional[float] = 1.0
10
prompt_template: Optional[str] = None
11
cache_mode: Optional[str] = "FP16"
12
draft: Optional['ModelCard'] = None
@@ -35,7 +35,7 @@ class ModelLoadRequest(BaseModel):
35
gpu_split_auto: Optional[bool] = True
36
gpu_split: Optional[List[float]] = Field(default_factory=list)
37
38
- rope_alpha: Optional[float] = 1.0
+ rope_alpha: Optional[float] = None
39
no_flash_attention: Optional[bool] = False
40
# low_mem: Optional[bool] = False
41
0 commit comments