Skip to content

Commit 368121a

Browse files
[DOCS] Adjusts note on minimum node size for ELSER and third-party models (#2447)
Co-authored-by: David Roberts <[email protected]>
1 parent 550b30a commit 368121a

File tree

2 files changed

+17
-3
lines changed

2 files changed

+17
-3
lines changed

docs/en/stack/ml/nlp/ml-nlp-elser.asciidoc

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -36,9 +36,14 @@ weights at index time, and to search against later.
3636
To use ELSER, you must have the {subscriptions}[appropriate subscription] level
3737
for semantic search or the trial period activated.
3838

39-
NOTE: The minimum dedicated ML node size for deploying and using
40-
the ELSER model is 4 GB. This is a minimum and better performance
41-
can be achieved by using bigger ML nodes.
39+
NOTE: The minimum dedicated ML node size for deploying and using the ELSER model
40+
is 4 GB in Elasticsearch Service if
41+
{cloud}/ec-autoscaling.html[deployment autoscaling] is turned off. Turning on
42+
autoscaling is recommended because it allows your deployment to dynamically
43+
adjust resources based on demand. Better performance can be achieved by using
44+
more allocations or more threads per allocation, which requires bigger ML nodes.
45+
Autoscaling provides bigger nodes when required. If autoscaling is turned off,
46+
you must provide suitably sized nodes yourself.
4247

4348

4449
[discrete]

docs/en/stack/ml/nlp/ml-nlp-model-ref.asciidoc

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,15 @@
99
:frontmatter-tags-content-type: [reference]
1010
:frontmatter-tags-user-goals: [analyze]
1111

12+
The minimum dedicated ML node size for deploying and using the ELSER model
13+
is 16 GB in Elasticsearch Service if
14+
{cloud}/ec-autoscaling.html[deployment autoscaling] is turned off. Turning on
15+
autoscaling is recommended because it allows your deployment to dynamically
16+
adjust resources based on demand. Better performance can be achieved by using
17+
more allocations or more threads per allocation, which requires bigger ML nodes.
18+
Autoscaling provides bigger nodes when required. If autoscaling is turned off,
19+
you must provide suitably sized nodes yourself.
20+
1221
The {stack-ml-features} support transformer models that conform to the standard
1322
BERT model interface and use the WordPiece tokenization algorithm.
1423

0 commit comments

Comments
 (0)