Skip to content

Commit

Permalink
Update serverless/pages/ml-nlp-auto-scale.mdx
Browse files Browse the repository at this point in the history
Co-authored-by: István Zoltán Szabó <istvan.szabo@elastic.co>
  • Loading branch information
kosabogi and szabosteve authored Oct 25, 2024
1 parent 26fb9e0 commit c8b8b3b
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion serverless/pages/ml-nlp-auto-scale.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,7 @@ The number of model allocations can be scaled down to 0.
They cannot be scaled up to more than 32 allocations, unless you explicitly set the maximum number of allocations to more.
Adaptive allocations must be set up independently for each deployment and [inference endpoint](https://www.elastic.co/guide/en/elasticsearch/reference/master/put-inference-api.html).

When you create inference endpoints on serverless deployments using Kibana, adaptive allocations are automatically turned on, and there is no option to disable them.
When you create inference endpoints on Serverless using Kibana, adaptive allocations are automatically turned on, and there is no option to disable them.

### Optimizing for typical use cases

Expand Down

0 comments on commit c8b8b3b

Please sign in to comment.