From 534f3a1ef34078947ae5587c1ee993ecd97b2a63 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Istv=C3=A1n=20Zolt=C3=A1n=20Szab=C3=B3?= Date: Tue, 29 Sep 2020 10:19:51 +0200 Subject: [PATCH] [DOCS] Adds DFA limitation item about number of training documents (#1370) --- .../stack/ml/df-analytics/ml-dfa-limitations.asciidoc | 11 +++++++++++ 1 file changed, 11 insertions(+) diff --git a/docs/en/stack/ml/df-analytics/ml-dfa-limitations.asciidoc b/docs/en/stack/ml/df-analytics/ml-dfa-limitations.asciidoc index c8ac3fa4c..f8691f07d 100644 --- a/docs/en/stack/ml/df-analytics/ml-dfa-limitations.asciidoc +++ b/docs/en/stack/ml/df-analytics/ml-dfa-limitations.asciidoc @@ -58,6 +58,17 @@ training percent. Run a few {dfanalytics-jobs} to see how the runtime scales with the increased number of data points and how the quality of results scales with an increased training percentage. + +[float] +[[dfa-training-docs]] +== {dfanalytics-jobs-cap} cannot use more than 2^32^ documents for training + +A {dfanalytics-job} that would use more than 2^32^ documents for training cannot +be started. The limitation applies only for documents participating in training +the model. If your source index contains more than 2^32^ documents, set the +`training_percent` to a value that represents less than 2^32^ documents. + + [float] [[dfa-missing-fields-limitations]] == Documents with missing values in analyzed fields are skipped