From 26fb9e08269d26278f6fbc32cd4650d3e3b2a548 Mon Sep 17 00:00:00 2001 From: kosabogi <105062005+kosabogi@users.noreply.github.com> Date: Fri, 25 Oct 2024 11:50:49 +0200 Subject: [PATCH] Update serverless/pages/ml-nlp-auto-scale.mdx MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Co-authored-by: István Zoltán Szabó --- serverless/pages/ml-nlp-auto-scale.mdx | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/serverless/pages/ml-nlp-auto-scale.mdx b/serverless/pages/ml-nlp-auto-scale.mdx index 7c9a7e0..38407ae 100644 --- a/serverless/pages/ml-nlp-auto-scale.mdx +++ b/serverless/pages/ml-nlp-auto-scale.mdx @@ -70,9 +70,9 @@ Increasing the number of threads will make the search processes more performant. You can enable adaptive resources for your models when starting or updating the model deployment. Adaptive resources make it possible for Elasticsearch to scale up or down the available resources based on the load on the process. This can help you to manage performance and cost more easily. -When adaptive resources are enabled, the number of vCPUs that the model deployment uses is set automatically based on the current load. -When the load is high, the number of vCPUs that the process can use is automatically increased. -When the load is low, the number of vCPUs that the process can use is automatically decreased. +When adaptive resources are enabled, the number of vCUs that the model deployment uses is set automatically based on the current load. +When the load is high, the number of vCUs that the process can use is automatically increased. +When the load is low, the number of vCUs that the process can use is automatically decreased. You can choose from three levels of resource usage for your trained model deployment; autoscaling will occur within the selected level's range.