You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -19,7 +19,7 @@ class ModelSettings(BaseSettings):
19
19
model: str=Field(
20
20
description="The path to the model to use for generating completions.",
21
21
)
22
-
model_alias: str|None=Field(
22
+
model_alias: Optional[str]=Field(
23
23
default=None,
24
24
description="The alias of the model to use for generating completions.",
25
25
)
@@ -38,7 +38,7 @@ class ModelSettings(BaseSettings):
38
38
ge=0,
39
39
description="Main GPU to use.",
40
40
)
41
-
tensor_split: list[float]|None=Field(
41
+
tensor_split: Optional[List[float]]=Field(
42
42
default=None,
43
43
description="Split layers across multiple GPUs in proportion.",
44
44
)
@@ -53,11 +53,11 @@ class ModelSettings(BaseSettings):
53
53
default=llama_cpp.llama_supports_mlock(),
54
54
description="Use mlock.",
55
55
)
56
-
kv_overrides: list[str]|None=Field(
56
+
kv_overrides: Optional[List[str]]=Field(
57
57
default=None,
58
58
description="List of model kv overrides in the format key=type:value where type is one of (bool, int, float). Valid true values are (true, TRUE, 1), otherwise false.",
59
59
)
60
-
rpc_servers: str|None=Field(
60
+
rpc_servers: Optional[str]=Field(
61
61
default=None,
62
62
description="comma seperated list of rpc servers for offloading",
63
63
)
@@ -109,25 +109,25 @@ class ModelSettings(BaseSettings):
109
109
description="Last n tokens to keep for repeat penalty calculation.",
110
110
)
111
111
# LoRA Params
112
-
lora_base: str|None=Field(
112
+
lora_base: Optional[str]=Field(
113
113
default=None,
114
114
description="Optional path to base model, useful if using a quantized base model and you want to apply LoRA to an f16 model.",
115
115
)
116
-
lora_path: str|None=Field(
116
+
lora_path: Optional[str]=Field(
117
117
default=None,
118
118
description="Path to a LoRA file to apply to the model.",
119
119
)
120
120
# Backend Params
121
-
numa: bool|int=Field(
121
+
numa: Union[bool, int]=Field(
122
122
default=False,
123
123
description="Enable NUMA support.",
124
124
)
125
125
# Chat Format Params
126
-
chat_format: str|None=Field(
126
+
chat_format: Optional[str]=Field(
127
127
default=None,
128
128
description="Chat format to use.",
129
129
)
130
-
clip_model_path: str|None=Field(
130
+
clip_model_path: Optional[str]=Field(
131
131
default=None,
132
132
description="Path to a CLIP model to use for multi-modal chat completion.",
133
133
)
@@ -145,21 +145,21 @@ class ModelSettings(BaseSettings):
145
145
description="The size of the cache in bytes. Only used if cache is True.",
146
146
)
147
147
# Tokenizer Options
148
-
hf_tokenizer_config_path: str|None=Field(
148
+
hf_tokenizer_config_path: Optional[str]=Field(
149
149
default=None,
150
150
description="The path to a HuggingFace tokenizer_config.json file.",
0 commit comments