File tree Expand file tree Collapse file tree 1 file changed +6
-1
lines changed Expand file tree Collapse file tree 1 file changed +6
-1
lines changed Original file line number Diff line number Diff line change @@ -124,14 +124,19 @@ PUT _inference/sparse_embedding/my-elser-model
124
124
{
125
125
"service": "elasticsearch",
126
126
"service_settings": {
127
- "num_allocations": 1,
127
+ "adaptive_allocations": {
128
+ "enabled": true,
129
+ "min_number_of_allocations": 1,
130
+ "max_number_of_allocations": 10
131
+ },
128
132
"num_threads": 1,
129
133
"model_id": ".elser_model_2_linux-x86_64"
130
134
}
131
135
}
132
136
----------------------------------
133
137
--
134
138
The API request automatically initiates the model download and then deploy the model.
139
+ This example uses <<ml-nlp-auto-scale,autoscaling>> through adaptive allocation.
135
140
136
141
Refer to the {ref}/infer-service-elser.html[ELSER {infer} service documentation] to learn more about the available settings.
137
142
You can’t perform that action at this time.
0 commit comments