From c8b8b3bcf4b876b9a7c07e3f1e79b7b1a8250577 Mon Sep 17 00:00:00 2001 From: kosabogi <105062005+kosabogi@users.noreply.github.com> Date: Fri, 25 Oct 2024 11:51:05 +0200 Subject: [PATCH] Update serverless/pages/ml-nlp-auto-scale.mdx MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Co-authored-by: István Zoltán Szabó --- serverless/pages/ml-nlp-auto-scale.mdx | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/serverless/pages/ml-nlp-auto-scale.mdx b/serverless/pages/ml-nlp-auto-scale.mdx index 38407ae..51c3284 100644 --- a/serverless/pages/ml-nlp-auto-scale.mdx +++ b/serverless/pages/ml-nlp-auto-scale.mdx @@ -53,7 +53,7 @@ The number of model allocations can be scaled down to 0. They cannot be scaled up to more than 32 allocations, unless you explicitly set the maximum number of allocations to more. Adaptive allocations must be set up independently for each deployment and [inference endpoint](https://www.elastic.co/guide/en/elasticsearch/reference/master/put-inference-api.html). -When you create inference endpoints on serverless deployments using Kibana, adaptive allocations are automatically turned on, and there is no option to disable them. +When you create inference endpoints on Serverless using Kibana, adaptive allocations are automatically turned on, and there is no option to disable them. ### Optimizing for typical use cases