From c6b3d1bfbf15b2f14ed44dc067c11511fcd489dd Mon Sep 17 00:00:00 2001 From: James Rodewig Date: Mon, 29 Apr 2019 10:33:15 -0400 Subject: [PATCH] [DOCS] Add anchors for Asciidoctor migration --- .../diversified-sampler-aggregation.asciidoc | 2 ++ .../bucket/filters-aggregation.asciidoc | 1 + .../bucket/sampler-aggregation.asciidoc | 1 + .../significantterms-aggregation.asciidoc | 1 + .../significanttext-aggregation.asciidoc | 4 ++-- .../metrics/weighted-avg-aggregation.asciidoc | 6 ++++++ .../pipeline/avg-bucket-aggregation.asciidoc | 3 +++ .../bucket-script-aggregation.asciidoc | 4 +++- .../bucket-selector-aggregation.asciidoc | 3 ++- .../pipeline/bucket-sort-aggregation.asciidoc | 2 ++ .../cumulative-sum-aggregation.asciidoc | 2 ++ .../pipeline/derivative-aggregation.asciidoc | 2 ++ ...extended-stats-bucket-aggregation.asciidoc | 2 ++ .../pipeline/max-bucket-aggregation.asciidoc | 2 ++ .../pipeline/min-bucket-aggregation.asciidoc | 2 ++ .../pipeline/movfn-aggregation.asciidoc | 20 +++++++++++++++++++ .../percentiles-bucket-aggregation.asciidoc | 2 ++ .../pipeline/serial-diff-aggregation.asciidoc | 2 ++ .../stats-bucket-aggregation.asciidoc | 2 ++ .../pipeline/sum-bucket-aggregation.asciidoc | 2 ++ .../synonym-graph-tokenfilter.asciidoc | 1 + .../tokenfilters/synonym-tokenfilter.asciidoc | 1 + docs/reference/docs/refresh.asciidoc | 1 + docs/reference/docs/update.asciidoc | 2 ++ docs/reference/getting-started.asciidoc | 1 + docs/reference/how-to/disk-usage.asciidoc | 3 +++ docs/reference/how-to/indexing-speed.asciidoc | 1 + docs/reference/how-to/search-speed.asciidoc | 2 ++ docs/reference/ilm/start-stop-ilm.asciidoc | 3 ++- .../ilm/using-policies-rollover.asciidoc | 2 ++ docs/reference/index-modules.asciidoc | 1 + .../reference/indices/rollover-index.asciidoc | 2 ++ docs/reference/indices/split-index.asciidoc | 1 + .../ingest/processors/dissect.asciidoc | 6 ++++++ .../mapping/dynamic/templates.asciidoc | 1 + .../mapping/fields/field-names-field.asciidoc | 1 + .../mapping/fields/source-field.asciidoc | 1 + .../mapping/params/fielddata.asciidoc | 2 ++ .../mapping/params/ignore-malformed.asciidoc | 1 + .../mapping/removal_of_types.asciidoc | 1 + .../mapping/types/geo-shape.asciidoc | 6 ++++++ docs/reference/mapping/types/ip.asciidoc | 1 + docs/reference/mapping/types/nested.asciidoc | 3 ++- .../mapping/types/percolator.asciidoc | 1 + .../migration/migrate_8_0/analysis.asciidoc | 1 + .../migration/migrate_8_0/security.asciidoc | 2 ++ .../modules/scripting/expression.asciidoc | 1 + docs/reference/modules/snapshots.asciidoc | 1 + docs/reference/modules/threadpool.asciidoc | 3 +++ docs/reference/query-dsl/bool-query.asciidoc | 1 + .../reference/query-dsl/exists-query.asciidoc | 2 ++ .../query-dsl/has-child-query.asciidoc | 1 + docs/reference/query-dsl/ids-query.asciidoc | 1 + .../query-dsl/multi-match-query.asciidoc | 3 +++ .../query-dsl/script-score-query.asciidoc | 9 ++++++--- .../rollup/rollup-search-limitations.asciidoc | 1 + docs/reference/search/profile.asciidoc | 5 ++++- docs/reference/search/rank-eval.asciidoc | 1 + docs/reference/settings/ssl-settings.asciidoc | 1 - docs/reference/setup/install/deb.asciidoc | 1 + docs/reference/setup/install/docker.asciidoc | 1 + docs/reference/setup/install/rpm.asciidoc | 1 + docs/reference/setup/starting.asciidoc | 4 ++++ docs/reference/sql/endpoints/jdbc.asciidoc | 2 ++ docs/reference/sql/limitations.asciidoc | 4 ++++ docs/reference/sql/security.asciidoc | 1 + 66 files changed, 146 insertions(+), 11 deletions(-) diff --git a/docs/reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc b/docs/reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc index 07d8261d200b4..edd6c6163da85 100644 --- a/docs/reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc +++ b/docs/reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc @@ -186,6 +186,7 @@ Please note that Elasticsearch will ignore the choice of execution hint if it is ==== Limitations +[[div-sampler-breadth-first-nested-agg]] ===== Cannot be nested under `breadth_first` aggregations Being a quality-based filter the diversified_sampler aggregation needs access to the relevance score produced for each document. It therefore cannot be nested under a `terms` aggregation which has the `collect_mode` switched from the default `depth_first` mode to `breadth_first` as this discards scores. @@ -194,6 +195,7 @@ In this situation an error will be thrown. ===== Limited de-dup logic. The de-duplication logic applies only at a shard level so will not apply across shards. +[[spec-syntax-geo-date-fields]] ===== No specialized syntax for geo/date fields Currently the syntax for defining the diversifying values is defined by a choice of `field` or `script` - there is no added syntactical sugar for expressing geo or date units such as "7d" (7 diff --git a/docs/reference/aggregations/bucket/filters-aggregation.asciidoc b/docs/reference/aggregations/bucket/filters-aggregation.asciidoc index 94b91654f0c7f..4e83c14a18ba7 100644 --- a/docs/reference/aggregations/bucket/filters-aggregation.asciidoc +++ b/docs/reference/aggregations/bucket/filters-aggregation.asciidoc @@ -118,6 +118,7 @@ request. The response for this example would be: // TESTRESPONSE[s/"_shards": \.\.\./"_shards": $body._shards/] // TESTRESPONSE[s/"hits": \.\.\./"hits": $body.hits/] +[[other-bucket]] ==== `Other` Bucket The `other_bucket` parameter can be set to add a bucket to the response which will contain all documents that do diff --git a/docs/reference/aggregations/bucket/sampler-aggregation.asciidoc b/docs/reference/aggregations/bucket/sampler-aggregation.asciidoc index c5ac91e9d3ad8..d3a255be1d5fe 100644 --- a/docs/reference/aggregations/bucket/sampler-aggregation.asciidoc +++ b/docs/reference/aggregations/bucket/sampler-aggregation.asciidoc @@ -155,6 +155,7 @@ The default value is 100. ==== Limitations +[[sampler-breadth-first-nested-agg]] ===== Cannot be nested under `breadth_first` aggregations Being a quality-based filter the sampler aggregation needs access to the relevance score produced for each document. It therefore cannot be nested under a `terms` aggregation which has the `collect_mode` switched from the default `depth_first` mode to `breadth_first` as this discards scores. diff --git a/docs/reference/aggregations/bucket/significantterms-aggregation.asciidoc b/docs/reference/aggregations/bucket/significantterms-aggregation.asciidoc index 5766cb1a73e58..50c238edb93c5 100644 --- a/docs/reference/aggregations/bucket/significantterms-aggregation.asciidoc +++ b/docs/reference/aggregations/bucket/significantterms-aggregation.asciidoc @@ -436,6 +436,7 @@ Available parameters in the script are `_subset_size`:: Number of documents in the subset. `_superset_size`:: Number of documents in the superset. +[[sig-terms-shard-size]] ===== Size & Shard Size The `size` parameter can be set to define how many term buckets should be returned out of the overall terms list. By diff --git a/docs/reference/aggregations/bucket/significanttext-aggregation.asciidoc b/docs/reference/aggregations/bucket/significanttext-aggregation.asciidoc index 429c822d3623d..3a9221c3c4372 100644 --- a/docs/reference/aggregations/bucket/significanttext-aggregation.asciidoc +++ b/docs/reference/aggregations/bucket/significanttext-aggregation.asciidoc @@ -92,7 +92,7 @@ It only occurs 5 times in our index as a whole (see the `bg_count`) and yet 4 of were lucky enough to appear in our 100 document sample of "bird flu" results. That suggests a significant word and one which the user can potentially add to their search. - +[[filter-duplicate-text-noisy-data]] ==== Dealing with noisy data using `filter_duplicate_text` Free-text fields often contain a mix of original content and mechanical copies of text (cut-and-paste biographies, email reply chains, retweets, boilerplate headers/footers, page navigation menus, sidebar news links, copyright notices, standard disclaimers, addresses). @@ -353,7 +353,7 @@ However, the `size` and `shard size` settings covered in the next section provid This aggregation supports the same scoring heuristics (JLH, mutual_information, gnd, chi_square etc) as the <> aggregation - +[[sig-text-shard-size]] ===== Size & Shard Size The `size` parameter can be set to define how many term buckets should be returned out of the overall terms list. By diff --git a/docs/reference/aggregations/metrics/weighted-avg-aggregation.asciidoc b/docs/reference/aggregations/metrics/weighted-avg-aggregation.asciidoc index 252728a6db367..70804a7904f1f 100644 --- a/docs/reference/aggregations/metrics/weighted-avg-aggregation.asciidoc +++ b/docs/reference/aggregations/metrics/weighted-avg-aggregation.asciidoc @@ -12,7 +12,9 @@ As a formula, a weighted average is the `∑(value * weight) / ∑(weight)` A regular average can be thought of as a weighted average where every value has an implicit weight of `1`. +[[weighted-avg-params]] .`weighted_avg` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`value` | The configuration for the field or script that provides the values |Required | @@ -23,7 +25,9 @@ A regular average can be thought of as a weighted average where every value has The `value` and `weight` objects have per-field specific configuration: +[[value-params]] .`value` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`field` | The field that values should be extracted from |Required | @@ -31,7 +35,9 @@ The `value` and `weight` objects have per-field specific configuration: |`script` | A script which provides the values for the document. This is mutually exclusive with `field` |Optional |=== +[[weight-params]] .`weight` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`field` | The field that weights should be extracted from |Required | diff --git a/docs/reference/aggregations/pipeline/avg-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/avg-bucket-aggregation.asciidoc index 274efcbce62fc..ea150b4ab6d8b 100644 --- a/docs/reference/aggregations/pipeline/avg-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/avg-bucket-aggregation.asciidoc @@ -4,6 +4,7 @@ A sibling pipeline aggregation which calculates the (mean) average value of a specified metric in a sibling aggregation. The specified metric must be numeric and the sibling aggregation must be a multi-bucket aggregation. +[[avg-bucket-agg-syntax]] ==== Syntax An `avg_bucket` aggregation looks like this in isolation: @@ -18,7 +19,9 @@ An `avg_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[avg-bucket-params]] .`avg_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the average for (see <> for more diff --git a/docs/reference/aggregations/pipeline/bucket-script-aggregation.asciidoc b/docs/reference/aggregations/pipeline/bucket-script-aggregation.asciidoc index 1825b37f0c734..7dc99ba7719cd 100644 --- a/docs/reference/aggregations/pipeline/bucket-script-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/bucket-script-aggregation.asciidoc @@ -4,6 +4,7 @@ A parent pipeline aggregation which executes a script which can perform per bucket computations on specified metrics in the parent multi-bucket aggregation. The specified metric must be numeric and the script must return a numeric value. +[[bucket-script-agg-syntax]] ==== Syntax A `bucket_script` aggregation looks like this in isolation: @@ -24,8 +25,9 @@ A `bucket_script` aggregation looks like this in isolation: <1> Here, `my_var1` is the name of the variable for this buckets path to use in the script, `the_sum` is the path to the metrics to use for that variable. - +[[bucket-script-params]] .`bucket_script` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`script` |The script to run for this aggregation. The script can be inline, file or indexed. (see <> diff --git a/docs/reference/aggregations/pipeline/bucket-selector-aggregation.asciidoc b/docs/reference/aggregations/pipeline/bucket-selector-aggregation.asciidoc index 4cc532c99c5d2..7ec19174a06e3 100644 --- a/docs/reference/aggregations/pipeline/bucket-selector-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/bucket-selector-aggregation.asciidoc @@ -29,8 +29,9 @@ A `bucket_selector` aggregation looks like this in isolation: <1> Here, `my_var1` is the name of the variable for this buckets path to use in the script, `the_sum` is the path to the metrics to use for that variable. - +[[bucket-selector-params]] .`bucket_selector` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`script` |The script to run for this aggregation. The script can be inline, file or indexed. (see <> diff --git a/docs/reference/aggregations/pipeline/bucket-sort-aggregation.asciidoc b/docs/reference/aggregations/pipeline/bucket-sort-aggregation.asciidoc index 633175dbf2825..a136a6ee4d578 100644 --- a/docs/reference/aggregations/pipeline/bucket-sort-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/bucket-sort-aggregation.asciidoc @@ -33,7 +33,9 @@ A `bucket_sort` aggregation looks like this in isolation: <1> Here, `sort_field_1` is the bucket path to the variable to be used as the primary sort and its order is ascending. +[[bucket-sort-params]] .`bucket_sort` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`sort` |The list of fields to sort on. See <> for more details. |Optional | diff --git a/docs/reference/aggregations/pipeline/cumulative-sum-aggregation.asciidoc b/docs/reference/aggregations/pipeline/cumulative-sum-aggregation.asciidoc index 748946f8bd671..a6dff7fa84651 100644 --- a/docs/reference/aggregations/pipeline/cumulative-sum-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/cumulative-sum-aggregation.asciidoc @@ -19,7 +19,9 @@ A `cumulative_sum` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[cumulative-sum-params]] .`cumulative_sum` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the cumulative sum for (see <> for more diff --git a/docs/reference/aggregations/pipeline/derivative-aggregation.asciidoc b/docs/reference/aggregations/pipeline/derivative-aggregation.asciidoc index 8479d1f45aea1..f40ace7432d57 100644 --- a/docs/reference/aggregations/pipeline/derivative-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/derivative-aggregation.asciidoc @@ -17,7 +17,9 @@ A `derivative` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[derivative-params]] .`derivative` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the derivative for (see <> for more diff --git a/docs/reference/aggregations/pipeline/extended-stats-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/extended-stats-bucket-aggregation.asciidoc index eeef705a6468d..c35223885fce0 100644 --- a/docs/reference/aggregations/pipeline/extended-stats-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/extended-stats-bucket-aggregation.asciidoc @@ -20,7 +20,9 @@ A `extended_stats_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[extended-stats-bucket-params]] .`extended_stats_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to calculate stats for (see <> for more diff --git a/docs/reference/aggregations/pipeline/max-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/max-bucket-aggregation.asciidoc index 8881315f50ab4..53a3aaa28f7e5 100644 --- a/docs/reference/aggregations/pipeline/max-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/max-bucket-aggregation.asciidoc @@ -19,7 +19,9 @@ A `max_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[max-bucket-params]] .`max_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the maximum for (see <> for more diff --git a/docs/reference/aggregations/pipeline/min-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/min-bucket-aggregation.asciidoc index ad6aaa28c90dd..620cf02c714fa 100644 --- a/docs/reference/aggregations/pipeline/min-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/min-bucket-aggregation.asciidoc @@ -19,7 +19,9 @@ A `min_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[min-bucket-params]] .`min_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the minimum for (see <> for more diff --git a/docs/reference/aggregations/pipeline/movfn-aggregation.asciidoc b/docs/reference/aggregations/pipeline/movfn-aggregation.asciidoc index 5745527bddd6f..f6fb25c76f662 100644 --- a/docs/reference/aggregations/pipeline/movfn-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/movfn-aggregation.asciidoc @@ -24,7 +24,9 @@ A `moving_fn` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[moving-avg-params]] .`moving_avg` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |Path to the metric of interest (see <> for more details |Required | @@ -188,7 +190,9 @@ The functions are available from the `MovingFunctions` namespace. E.g. `MovingF This function accepts a collection of doubles and returns the maximum value in that window. `null` and `NaN` values are ignored; the maximum is only calculated over the real values. If the window is empty, or all values are `null`/`NaN`, `NaN` is returned as the result. +[[max-params]] .`max(double[] values)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the maximum @@ -229,7 +233,9 @@ POST /_search This function accepts a collection of doubles and returns the minimum value in that window. `null` and `NaN` values are ignored; the minimum is only calculated over the real values. If the window is empty, or all values are `null`/`NaN`, `NaN` is returned as the result. +[[min-params]] .`min(double[] values)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the minimum @@ -270,7 +276,9 @@ POST /_search This function accepts a collection of doubles and returns the sum of the values in that window. `null` and `NaN` values are ignored; the sum is only calculated over the real values. If the window is empty, or all values are `null`/`NaN`, `0.0` is returned as the result. +[[sum-params]] .`sum(double[] values)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the sum of @@ -312,7 +320,9 @@ This function accepts a collection of doubles and average, then returns the stan `null` and `NaN` values are ignored; the sum is only calculated over the real values. If the window is empty, or all values are `null`/`NaN`, `0.0` is returned as the result. +[[stddev-params]] .`stdDev(double[] values)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the standard deviation of @@ -363,7 +373,9 @@ the values from a `simple` moving average tend to "lag" behind the real data. `null`/`NaN`, `NaN` is returned as the result. This means that the count used in the average calculation is count of non-`null`,non-`NaN` values. +[[unweightedavg-params]] .`unweightedAvg(double[] values)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the sum of @@ -407,7 +419,9 @@ the "lag" behind the data's mean, since older points have less influence. If the window is empty, or all values are `null`/`NaN`, `NaN` is returned as the result. +[[linearweightedavg-params]] .`linearWeightedAvg(double[] values)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the sum of @@ -456,7 +470,9 @@ moving average. This tends to make the moving average track the data more close `null`/`NaN`, `NaN` is returned as the result. This means that the count used in the average calculation is count of non-`null`,non-`NaN` values. +[[ewma-params]] .`ewma(double[] values, double alpha)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the sum of @@ -511,7 +527,9 @@ Values are produced by multiplying the level and trend components. `null`/`NaN`, `NaN` is returned as the result. This means that the count used in the average calculation is count of non-`null`,non-`NaN` values. +[[holt-params]] .`holt(double[] values, double alpha)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the sum of @@ -572,7 +590,9 @@ for future enhancements. `null`/`NaN`, `NaN` is returned as the result. This means that the count used in the average calculation is count of non-`null`,non-`NaN` values. +[[holtwinters-params]] .`holtWinters(double[] values, double alpha)` Parameters +[options="header"] |=== |Parameter Name |Description |`values` |The window of values to find the sum of diff --git a/docs/reference/aggregations/pipeline/percentiles-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/percentiles-bucket-aggregation.asciidoc index 032b6ef4e419d..456a4046c0624 100644 --- a/docs/reference/aggregations/pipeline/percentiles-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/percentiles-bucket-aggregation.asciidoc @@ -18,7 +18,9 @@ A `percentiles_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[percentiles-bucket-params]] .`percentiles_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the percentiles for (see <> for more diff --git a/docs/reference/aggregations/pipeline/serial-diff-aggregation.asciidoc b/docs/reference/aggregations/pipeline/serial-diff-aggregation.asciidoc index 70aea68f88c34..1506e39685845 100644 --- a/docs/reference/aggregations/pipeline/serial-diff-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/serial-diff-aggregation.asciidoc @@ -46,7 +46,9 @@ A `serial_diff` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[serial-diff-params]] .`serial_diff` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |Path to the metric of interest (see <> for more details |Required | diff --git a/docs/reference/aggregations/pipeline/stats-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/stats-bucket-aggregation.asciidoc index b9c52ae981f75..2c9f585ebea3a 100644 --- a/docs/reference/aggregations/pipeline/stats-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/stats-bucket-aggregation.asciidoc @@ -18,7 +18,9 @@ A `stats_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[stats-bucket-params]] .`stats_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to calculate stats for (see <> for more diff --git a/docs/reference/aggregations/pipeline/sum-bucket-aggregation.asciidoc b/docs/reference/aggregations/pipeline/sum-bucket-aggregation.asciidoc index b39cf472323c2..83e0e32135040 100644 --- a/docs/reference/aggregations/pipeline/sum-bucket-aggregation.asciidoc +++ b/docs/reference/aggregations/pipeline/sum-bucket-aggregation.asciidoc @@ -18,7 +18,9 @@ A `sum_bucket` aggregation looks like this in isolation: -------------------------------------------------- // NOTCONSOLE +[[sum-bucket-params]] .`sum_bucket` Parameters +[options="header"] |=== |Parameter Name |Description |Required |Default Value |`buckets_path` |The path to the buckets we wish to find the sum for (see <> for more diff --git a/docs/reference/analysis/tokenfilters/synonym-graph-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/synonym-graph-tokenfilter.asciidoc index 785eb77f2c65e..2285c6f6e8989 100644 --- a/docs/reference/analysis/tokenfilters/synonym-graph-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/synonym-graph-tokenfilter.asciidoc @@ -93,6 +93,7 @@ set to `false` no mapping would get added as when `expand=false` the target mapp stop word. [float] +[[synonym-graph-tokenizer-ignore_case-deprecated]] ==== `tokenizer` and `ignore_case` are deprecated The `tokenizer` parameter controls the tokenizers that will be used to diff --git a/docs/reference/analysis/tokenfilters/synonym-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/synonym-tokenfilter.asciidoc index 87c99f6f38683..1107236194655 100644 --- a/docs/reference/analysis/tokenfilters/synonym-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/synonym-tokenfilter.asciidoc @@ -83,6 +83,7 @@ stop word. [float] +[[synonym-tokenizer-ignore_case-deprecated]] ==== `tokenizer` and `ignore_case` are deprecated The `tokenizer` parameter controls the tokenizers that will be used to diff --git a/docs/reference/docs/refresh.asciidoc b/docs/reference/docs/refresh.asciidoc index e5051497ecda3..121e0c494d828 100644 --- a/docs/reference/docs/refresh.asciidoc +++ b/docs/reference/docs/refresh.asciidoc @@ -63,6 +63,7 @@ general, if you have a running system you don't wish to disturb then `refresh=wait_for` is a smaller modification. [float] +[[refresh_wait_for-force-refresh]] === `refresh=wait_for` Can Force a Refresh If a `refresh=wait_for` request comes in when there are already diff --git a/docs/reference/docs/update.asciidoc b/docs/reference/docs/update.asciidoc index 66e5783e7eb9b..00cd66232190f 100644 --- a/docs/reference/docs/update.asciidoc +++ b/docs/reference/docs/update.asciidoc @@ -243,6 +243,7 @@ POST test/_update/1 // TEST[continued] [float] +[[scripted_upsert]] ==== `scripted_upsert` If you would like your script to run regardless of whether the document exists @@ -272,6 +273,7 @@ POST sessions/_update/dh3sgudg8gsrgl // TEST[continued] [float] +[[doc_as_upsert]] ==== `doc_as_upsert` Instead of sending a partial `doc` plus an `upsert` doc, setting diff --git a/docs/reference/getting-started.asciidoc b/docs/reference/getting-started.asciidoc index 6e112559f8771..b81d2b284371d 100755 --- a/docs/reference/getting-started.asciidoc +++ b/docs/reference/getting-started.asciidoc @@ -505,6 +505,7 @@ This REST access pattern is so pervasive throughout all the API commands that if Elasticsearch provides data manipulation and search capabilities in near real time. By default, you can expect a one second delay (refresh interval) from the time you index/update/delete your data until the time that it appears in your search results. This is an important distinction from other platforms like SQL wherein data is immediately available after a transaction is completed. [float] +[[indexing-replacing-documents]] === Indexing/Replacing Documents We've previously seen how we can index a single document. Let's recall that command again: diff --git a/docs/reference/how-to/disk-usage.asciidoc b/docs/reference/how-to/disk-usage.asciidoc index 713e03d188ad9..a475122b99ec2 100644 --- a/docs/reference/how-to/disk-usage.asciidoc +++ b/docs/reference/how-to/disk-usage.asciidoc @@ -91,6 +91,7 @@ PUT index // CONSOLE [float] +[[default-dynamic-string-mapping]] === Don't use default dynamic string mappings The default <> will index string fields @@ -133,11 +134,13 @@ Larger shards are going to be more efficient at storing data. To increase the si Keep in mind that large shard sizes come with drawbacks, such as long full recovery times. [float] +[[disable-source]] === Disable `_source` The <> field stores the original JSON body of the document. If you don’t need access to it you can disable it. However, APIs that needs access to `_source` such as update and reindex won’t work. [float] +[[best-compression]] === Use `best_compression` The `_source` and stored fields can easily take a non negligible amount of disk diff --git a/docs/reference/how-to/indexing-speed.asciidoc b/docs/reference/how-to/indexing-speed.asciidoc index 1c8b989779d6c..147d3a557c830 100644 --- a/docs/reference/how-to/indexing-speed.asciidoc +++ b/docs/reference/how-to/indexing-speed.asciidoc @@ -17,6 +17,7 @@ it is advisable to avoid going beyond a couple tens of megabytes per request even if larger requests seem to perform better. [float] +[[multiple-workers-threads]] === Use multiple workers/threads to send data to Elasticsearch A single thread sending bulk requests is unlikely to be able to max out the diff --git a/docs/reference/how-to/search-speed.asciidoc b/docs/reference/how-to/search-speed.asciidoc index 017dd74afa084..e09b18df20c5d 100644 --- a/docs/reference/how-to/search-speed.asciidoc +++ b/docs/reference/how-to/search-speed.asciidoc @@ -161,6 +161,7 @@ GET index/_search // TEST[continued] [float] +[[map-ids-as-keyword]] === Consider mapping identifiers as `keyword` The fact that some data is numeric does not mean it should always be mapped as a @@ -354,6 +355,7 @@ conjunctions faster at the cost of slightly slower indexing. Read more about it in the <>. [float] +[[preference-cache-optimization]] === Use `preference` to optimize cache utilization There are multiple caches that can help with search performance, such as the diff --git a/docs/reference/ilm/start-stop-ilm.asciidoc b/docs/reference/ilm/start-stop-ilm.asciidoc index e5366f028a9c7..22ca0ae48fd98 100644 --- a/docs/reference/ilm/start-stop-ilm.asciidoc +++ b/docs/reference/ilm/start-stop-ilm.asciidoc @@ -68,8 +68,9 @@ If the request does not encounter errors, you receive the following result: The operating modes of ILM: - +[[ilm-operating-modes]] .ILM Operating Modes +[options="header"] |=== |Name |Description |RUNNING |Normal operation where all policies are executed as normal diff --git a/docs/reference/ilm/using-policies-rollover.asciidoc b/docs/reference/ilm/using-policies-rollover.asciidoc index dfb89eb26d121..11be1bed85a16 100644 --- a/docs/reference/ilm/using-policies-rollover.asciidoc +++ b/docs/reference/ilm/using-policies-rollover.asciidoc @@ -28,7 +28,9 @@ new index. The rollover action takes the following parameters: +[[rollover-action-params]] .`rollover` Action Parameters +[options="header"] |=== |Name |Description |max_size |The maximum estimated size the primary shard of the index is allowed diff --git a/docs/reference/index-modules.asciidoc b/docs/reference/index-modules.asciidoc index 96ccc9ab6a8a4..ef7c40b11d048 100644 --- a/docs/reference/index-modules.asciidoc +++ b/docs/reference/index-modules.asciidoc @@ -277,6 +277,7 @@ Other index settings are available in index modules: Control over the transaction log and background flush operations. [float] +[[x-pack-index-settings]] === [xpack]#{xpack} index settings# <>:: diff --git a/docs/reference/indices/rollover-index.asciidoc b/docs/reference/indices/rollover-index.asciidoc index 7f6c3d0452a2d..b5037d0a94233 100644 --- a/docs/reference/indices/rollover-index.asciidoc +++ b/docs/reference/indices/rollover-index.asciidoc @@ -26,7 +26,9 @@ will now have the rollover alias pointing to it as the write index with `is_writ The available conditions are: +[[index-rollover-conditions]] .`conditions` parameters +[options="header"] |=== | Name | Description | max_age | The maximum age of the index diff --git a/docs/reference/indices/split-index.asciidoc b/docs/reference/indices/split-index.asciidoc index d79c0b0e9e250..4b9aaa09a94b6 100644 --- a/docs/reference/indices/split-index.asciidoc +++ b/docs/reference/indices/split-index.asciidoc @@ -50,6 +50,7 @@ Splitting works as follows: had just been re-opened. [float] +[[incremental-resharding]] === Why doesn't Elasticsearch support incremental resharding? Going from `N` shards to `N+1` shards, aka. incremental resharding, is indeed a diff --git a/docs/reference/ingest/processors/dissect.asciidoc b/docs/reference/ingest/processors/dissect.asciidoc index aa54cb4411f3e..f2a7a42b9d668 100644 --- a/docs/reference/ingest/processors/dissect.asciidoc +++ b/docs/reference/ingest/processors/dissect.asciidoc @@ -78,12 +78,14 @@ include::common-options.asciidoc[] } -------------------------------------------------- // NOTCONSOLE + [[dissect-key-modifiers]] ==== Dissect key modifiers Key modifiers can change the default behavior for dissection. Key modifiers may be found on the left or right of the `%{keyname}` always inside the `%{` and `}`. For example `%{+keyname ->}` has the append and right padding modifiers. +[[dissect-key-modifiers-table]] .Dissect Key Modifiers [options="header"] |====== @@ -132,6 +134,7 @@ Right padding modifier with empty key example * level = WARN |====== +[[append-modifier]] ===== Append modifier (`+`) [[dissect-modifier-append-key]] Dissect supports appending two or more results together for the output. @@ -146,6 +149,7 @@ Append modifier example * name = john jacob jingleheimer schmidt |====== +[[append-order-modifier]] ===== Append with order modifier (`+` and `/n`) [[dissect-modifier-append-key-with-order]] Dissect supports appending two or more results together for the output. @@ -160,6 +164,7 @@ Append with order modifier example * name = schmidt,john,jingleheimer,jacob |====== +[[named-skip-key]] ===== Named skip key (`?`) [[dissect-modifier-named-skip-key]] Dissect supports ignoring matches in the final result. This can be done with an empty key `%{}`, but for readability @@ -174,6 +179,7 @@ Named skip key modifier example * @timestamp = 30/Apr/1998:22:00:52 +0000 |====== +[[reference-keys]] ===== Reference keys (`*` and `&`) [[dissect-modifier-reference-keys]] Dissect support using parsed values as the key/value pairings for the structured content. Imagine a system that diff --git a/docs/reference/mapping/dynamic/templates.asciidoc b/docs/reference/mapping/dynamic/templates.asciidoc index 8598eab412e79..a55fd98d91e67 100644 --- a/docs/reference/mapping/dynamic/templates.asciidoc +++ b/docs/reference/mapping/dynamic/templates.asciidoc @@ -286,6 +286,7 @@ PUT my_index -------------------------------------------------- // CONSOLE +[[text-only-mappings-strings]] ===== `text`-only mappings for strings On the contrary to the previous example, if the only thing that you care about diff --git a/docs/reference/mapping/fields/field-names-field.asciidoc b/docs/reference/mapping/fields/field-names-field.asciidoc index c455c55f5ea7f..1ae4ab4c8fc89 100644 --- a/docs/reference/mapping/fields/field-names-field.asciidoc +++ b/docs/reference/mapping/fields/field-names-field.asciidoc @@ -11,6 +11,7 @@ Now the `_field_names` field only indexes the names of fields that have or `norm` enabled the <> query will still be available but will not use the `_field_names` field. +[[disable-field-names]] ==== Disabling `_field_names` Disabling `_field_names` is often not necessary because it no longer diff --git a/docs/reference/mapping/fields/source-field.asciidoc b/docs/reference/mapping/fields/source-field.asciidoc index 757fc0fa5b662..c9fd2cf186909 100644 --- a/docs/reference/mapping/fields/source-field.asciidoc +++ b/docs/reference/mapping/fields/source-field.asciidoc @@ -6,6 +6,7 @@ at index time. The `_source` field itself is not indexed (and thus is not searchable), but it is stored so that it can be returned when executing _fetch_ requests, like <> or <>. +[[disable-source-field]] ==== Disabling the `_source` field Though very handy to have around, the source field does incur storage overhead diff --git a/docs/reference/mapping/params/fielddata.asciidoc b/docs/reference/mapping/params/fielddata.asciidoc index 42f02b7ee28ea..8e46dd0370245 100644 --- a/docs/reference/mapping/params/fielddata.asciidoc +++ b/docs/reference/mapping/params/fielddata.asciidoc @@ -19,6 +19,7 @@ reading the entire inverted index for each segment from disk, inverting the term ↔︎ document relationship, and storing the result in memory, in the JVM heap. +[[fielddata-disabled-text-fields]] ==== Fielddata is disabled on `text` fields by default Fielddata can consume a *lot* of heap space, especially when loading high @@ -75,6 +76,7 @@ PUT my_index <1> Use the `my_field` field for searches. <2> Use the `my_field.keyword` field for aggregations, sorting, or in scripts. +[[enable-fielddata-text-fields]] ==== Enabling fielddata on `text` fields You can enable fielddata on an existing `text` field using the diff --git a/docs/reference/mapping/params/ignore-malformed.asciidoc b/docs/reference/mapping/params/ignore-malformed.asciidoc index f8f9a4591f947..8c91bb48ee7cd 100644 --- a/docs/reference/mapping/params/ignore-malformed.asciidoc +++ b/docs/reference/mapping/params/ignore-malformed.asciidoc @@ -91,6 +91,7 @@ become meaningless. Elasticsearch makes it easy to check how many documents have malformed fields by using `exist` or `term` queries on the special <> field. +[[json-object-limits]] ==== Limits for JSON Objects You can't use `ignore_malformed` with the following datatypes: diff --git a/docs/reference/mapping/removal_of_types.asciidoc b/docs/reference/mapping/removal_of_types.asciidoc index 8e78f3f140596..4d26f30ada256 100644 --- a/docs/reference/mapping/removal_of_types.asciidoc +++ b/docs/reference/mapping/removal_of_types.asciidoc @@ -216,6 +216,7 @@ GET twitter/_search <1> The explicit `type` field takes the place of the implicit `_type` field. [float] +[[parent-child-mapping-types]] ==== Parent/Child without mapping types Previously, a parent-child relationship was represented by making one mapping diff --git a/docs/reference/mapping/types/geo-shape.asciidoc b/docs/reference/mapping/types/geo-shape.asciidoc index a46b8a3f8a87c..1cf85e305a95d 100644 --- a/docs/reference/mapping/types/geo-shape.asciidoc +++ b/docs/reference/mapping/types/geo-shape.asciidoc @@ -323,6 +323,7 @@ POST /example/_doc // CONSOLE [float] +[[linestring]] ===== http://geojson.org/geojson-spec.html#id3[LineString] A `linestring` defined by an array of two or more positions. By @@ -357,6 +358,7 @@ The above `linestring` would draw a straight line starting at the White House to the US Capitol Building. [float] +[[polygon]] ===== http://www.geojson.org/geojson-spec.html#id4[Polygon] A polygon is defined by a list of a list of points. The first and last @@ -473,6 +475,7 @@ POST /example/_doc // CONSOLE [float] +[[multipoint]] ===== http://www.geojson.org/geojson-spec.html#id5[MultiPoint] The following is an example of a list of geojson points: @@ -503,6 +506,7 @@ POST /example/_doc // CONSOLE [float] +[[multilinestring]] ===== http://www.geojson.org/geojson-spec.html#id6[MultiLineString] The following is an example of a list of geojson linestrings: @@ -535,6 +539,7 @@ POST /example/_doc // CONSOLE [float] +[[multipolygon]] ===== http://www.geojson.org/geojson-spec.html#id7[MultiPolygon] The following is an example of a list of geojson polygons (second polygon contains a hole): @@ -567,6 +572,7 @@ POST /example/_doc // CONSOLE [float] +[[geometry_collection]] ===== http://geojson.org/geojson-spec.html#geometrycollection[Geometry Collection] The following is an example of a collection of geojson geometry objects: diff --git a/docs/reference/mapping/types/ip.asciidoc b/docs/reference/mapping/types/ip.asciidoc index 7f3f5f57d7077..46ca444c9558d 100644 --- a/docs/reference/mapping/types/ip.asciidoc +++ b/docs/reference/mapping/types/ip.asciidoc @@ -69,6 +69,7 @@ The following parameters are accepted by `ip` fields: the <> field. Accepts `true` or `false` (default). +[[query-ip-fields]] ==== Querying `ip` fields The most common way to query ip addresses is to use the diff --git a/docs/reference/mapping/types/nested.asciidoc b/docs/reference/mapping/types/nested.asciidoc index f420e680c8590..5969dcfd6956b 100644 --- a/docs/reference/mapping/types/nested.asciidoc +++ b/docs/reference/mapping/types/nested.asciidoc @@ -65,6 +65,7 @@ GET my_index/_search // CONSOLE // TEST[continued] +[[nested-fields-array-objects]] ==== Using `nested` fields for arrays of objects If you need to index arrays of objects and to maintain the independence of @@ -200,7 +201,7 @@ document is indexed as a separate document. To safeguard against ill-defined map the number of nested fields that can be defined per index has been limited to 50. See <>. - +[[limit-nested-json-objects-number]] ==== Limiting the number of `nested` json objects Indexing a document with an array of 100 objects within a nested field, will actually create 101 documents, as each nested object will be indexed as a separate document. diff --git a/docs/reference/mapping/types/percolator.asciidoc b/docs/reference/mapping/types/percolator.asciidoc index ce840e1379de3..00270fff38538 100644 --- a/docs/reference/mapping/types/percolator.asciidoc +++ b/docs/reference/mapping/types/percolator.asciidoc @@ -720,6 +720,7 @@ fail. ==== Limitations [float] +[[parent-child]] ===== Parent/child Because the `percolate` query is processing one document at a time, it doesn't support queries and filters that run diff --git a/docs/reference/migration/migrate_8_0/analysis.asciidoc b/docs/reference/migration/migrate_8_0/analysis.asciidoc index f9a4bb76fd841..dd4b4ae73e920 100644 --- a/docs/reference/migration/migrate_8_0/analysis.asciidoc +++ b/docs/reference/migration/migrate_8_0/analysis.asciidoc @@ -10,6 +10,7 @@ // end::notable-breaking-changes[] [float] +[[ngram-edgengram-filter-names-removed]] ==== The `nGram` and `edgeNGram` token filter names have been removed The `nGram` and `edgeNGram` token filter names that have been deprecated since diff --git a/docs/reference/migration/migrate_8_0/security.asciidoc b/docs/reference/migration/migrate_8_0/security.asciidoc index e4caca1e8fe43..fcc0a5b22168a 100644 --- a/docs/reference/migration/migrate_8_0/security.asciidoc +++ b/docs/reference/migration/migrate_8_0/security.asciidoc @@ -10,12 +10,14 @@ // end::notable-breaking-changes[] [float] +[[accept-default-password-removed]] ==== The `accept_default_password` setting has been removed The `xpack.security.authc.accept_default_password` setting has not had any affect since the 6.0 release of {es}. It has been removed and cannot be used. [float] +[[roles-index-cache-removed]] ==== The `roles.index.cache.*` settings have been removed The `xpack.security.authz.store.roles.index.cache.max_size` and diff --git a/docs/reference/modules/scripting/expression.asciidoc b/docs/reference/modules/scripting/expression.asciidoc index fb7739261ef4a..de3704a899e52 100644 --- a/docs/reference/modules/scripting/expression.asciidoc +++ b/docs/reference/modules/scripting/expression.asciidoc @@ -112,6 +112,7 @@ The following example shows the difference in years between the `date` fields da `doc['date1'].date.year - doc['date0'].date.year` [float] +[[geo-point-field-api]] === `geo_point` field API [cols="<,<",options="header",] |======================================================================= diff --git a/docs/reference/modules/snapshots.asciidoc b/docs/reference/modules/snapshots.asciidoc index b206b72fb31a9..ec7916d5a3445 100644 --- a/docs/reference/modules/snapshots.asciidoc +++ b/docs/reference/modules/snapshots.asciidoc @@ -694,6 +694,7 @@ GET /_snapshot/my_backup/snapshot_1,snapshot_2/_status // TEST[continued] [float] +[[monitor-snapshot-restore-progress]] === Monitoring snapshot/restore progress There are several ways to monitor the progress of the snapshot and restores processes while they are running. Both diff --git a/docs/reference/modules/threadpool.asciidoc b/docs/reference/modules/threadpool.asciidoc index 38f8d40e67f23..d0f68e37730b8 100644 --- a/docs/reference/modules/threadpool.asciidoc +++ b/docs/reference/modules/threadpool.asciidoc @@ -69,6 +69,7 @@ thread_pool: The following are the types of thread pools and their respective parameters: [float] +[[fixed]] ==== `fixed` The `fixed` thread pool holds a fixed size of threads to handle the @@ -92,6 +93,7 @@ thread_pool: -------------------------------------------------- [float] +[[fixed-auto-queue-size]] ==== `fixed_auto_queue_size` experimental[] @@ -138,6 +140,7 @@ thread_pool: -------------------------------------------------- [float] +[[scaling]] ==== `scaling` The `scaling` thread pool holds a dynamic number of threads. This diff --git a/docs/reference/query-dsl/bool-query.asciidoc b/docs/reference/query-dsl/bool-query.asciidoc index d4b5919454836..49dc4e2364efb 100644 --- a/docs/reference/query-dsl/bool-query.asciidoc +++ b/docs/reference/query-dsl/bool-query.asciidoc @@ -58,6 +58,7 @@ POST _search -------------------------------------------------- // CONSOLE +[[score-bool-filter]] ==== Scoring with `bool.filter` Queries specified under the `filter` element have no effect on scoring -- diff --git a/docs/reference/query-dsl/exists-query.asciidoc b/docs/reference/query-dsl/exists-query.asciidoc index b2e27c76e494d..dc624c19039ef 100644 --- a/docs/reference/query-dsl/exists-query.asciidoc +++ b/docs/reference/query-dsl/exists-query.asciidoc @@ -44,6 +44,7 @@ These documents would *not* match the above query: <3> The `user` field is missing completely. [float] +[[null-value-mapping]] ==== `null_value` mapping If the field mapping includes the <> setting @@ -86,6 +87,7 @@ no values in the `user` field and thus would not match the `exists` filter: -------------------------------------------------- // NOTCONSOLE +[[missing-query]] ==== `missing` query There isn't a `missing` query. Instead use the `exists` query inside a diff --git a/docs/reference/query-dsl/has-child-query.asciidoc b/docs/reference/query-dsl/has-child-query.asciidoc index ca23afb0fae26..b75c06c99c566 100644 --- a/docs/reference/query-dsl/has-child-query.asciidoc +++ b/docs/reference/query-dsl/has-child-query.asciidoc @@ -62,6 +62,7 @@ GET /_search // CONSOLE [float] +[[min-max-children]] ==== Min/Max Children The `has_child` query allows you to specify that a minimum and/or maximum diff --git a/docs/reference/query-dsl/ids-query.asciidoc b/docs/reference/query-dsl/ids-query.asciidoc index 6056c24702c91..70554e1acbf1b 100644 --- a/docs/reference/query-dsl/ids-query.asciidoc +++ b/docs/reference/query-dsl/ids-query.asciidoc @@ -18,6 +18,7 @@ GET /_search -------------------------------------------------- // CONSOLE +[[ids-query-top-level-parameters]] ==== Top-level parameters for `ids` [cols="v,v",options="header"] diff --git a/docs/reference/query-dsl/multi-match-query.asciidoc b/docs/reference/query-dsl/multi-match-query.asciidoc index 703abd42ee621..9f574ed814d3c 100644 --- a/docs/reference/query-dsl/multi-match-query.asciidoc +++ b/docs/reference/query-dsl/multi-match-query.asciidoc @@ -21,6 +21,7 @@ GET /_search <2> The fields to be queried. [float] +[[field-boost]] ==== `fields` and per-field boosting Fields can be specified with wildcards, eg: @@ -391,6 +392,7 @@ Also, accepts `analyzer`, `boost`, `operator`, `minimum_should_match`, `lenient`, `zero_terms_query` and `cutoff_frequency`, as explained in <>. +[[cross-field-analysis]] ===== `cross_field` and analysis The `cross_field` type can only work in term-centric mode on fields that have @@ -499,6 +501,7 @@ which will be executed as: blended("will", fields: [first, first.edge, last.edge, last]) blended("smith", fields: [first, first.edge, last.edge, last]) +[[tie-breaker]] ===== `tie_breaker` By default, each per-term `blended` query will use the best score returned by diff --git a/docs/reference/query-dsl/script-score-query.asciidoc b/docs/reference/query-dsl/script-score-query.asciidoc index e6418d408d7f2..42d4a7b1517e3 100644 --- a/docs/reference/query-dsl/script-score-query.asciidoc +++ b/docs/reference/query-dsl/script-score-query.asciidoc @@ -233,7 +233,7 @@ field, whose only drawback is that scores will change if the document is updated since update operations also update the value of the `_seq_no` field. -[[decay-functions]] +[[decay-functions-numeric-fields]] ===== Decay functions for numeric fields You can read more about decay functions {ref}/query-dsl-function-score-query.html#function-decay[here]. @@ -323,10 +323,12 @@ Script Score Query will be a substitute for it. Here we describe how Function Score Query's functions can be equivalently implemented in Script Score Query: +[[script-score]] ===== `script_score` What you used in `script_score` of the Function Score query, you can copy into the Script Score query. No changes here. +[[weight]] ===== `weight` `weight` function can be implemented in the Script Score query through the following script: @@ -342,12 +344,13 @@ the following script: -------------------------------------------------- // NOTCONSOLE +[[random-score]] ===== `random_score` Use `randomScore` function as described in <>. - +[[field-value-factor]] ===== `field_value_factor` `field_value_factor` function can be easily implemented through script: @@ -397,7 +400,7 @@ through a script: | `reciprocal` | `1.0 / doc['f'].value` |======================================================================= - +[[decay-functions]] ===== `decay functions` Script Score query has equivalent <> that can be used in script. diff --git a/docs/reference/rollup/rollup-search-limitations.asciidoc b/docs/reference/rollup/rollup-search-limitations.asciidoc index c8a736450bde0..81cb162bd3c11 100644 --- a/docs/reference/rollup/rollup-search-limitations.asciidoc +++ b/docs/reference/rollup/rollup-search-limitations.asciidoc @@ -32,6 +32,7 @@ To help simplify the problem, we have limited search to just one rollup index at may be able to open this up to multiple rollup jobs. [float] +[[aggregate-stored-only]] === Can only aggregate what's been stored A perhaps obvious limitation, but rollups can only aggregate on data that has been stored in the rollups. If you don't configure the diff --git a/docs/reference/search/profile.asciidoc b/docs/reference/search/profile.asciidoc index 9a6779defab39..e5ba1b851cdc8 100644 --- a/docs/reference/search/profile.asciidoc +++ b/docs/reference/search/profile.asciidoc @@ -243,6 +243,7 @@ sufficient to see that a particular component of a query is slow, and not necess the `advance` phase of that query is the cause, for example. ======================================= +[[query-section]] ==== `query` Section The `query` section contains detailed timing of the query tree executed by Lucene on a particular shard. @@ -399,6 +400,7 @@ The meaning of the stats are as follows: means the `nextDoc()` method was called on two different documents. This can be used to help judge how selective queries are, by comparing counts between different query components. +[[collectors-section]] ==== `collectors` Section The Collectors portion of the response shows high-level execution details. Lucene works by defining a "Collector" @@ -485,7 +487,7 @@ For reference, the various collector reasons are: match_all query (which you will see added to the Query section) to collect your entire dataset - +[[rewrite-section]] ==== `rewrite` Section All queries in Lucene undergo a "rewriting" process. A query (and its sub-queries) may be rewritten one or @@ -694,6 +696,7 @@ Hopefully this will be fixed in future iterations, but it is a tricky problem to [[search-profile-aggregations]] === Profiling Aggregations +[[agg-section]] ==== `aggregations` Section diff --git a/docs/reference/search/rank-eval.asciidoc b/docs/reference/search/rank-eval.asciidoc index 29176f15f8cd4..b25048fec90e6 100644 --- a/docs/reference/search/rank-eval.asciidoc +++ b/docs/reference/search/rank-eval.asciidoc @@ -136,6 +136,7 @@ The `metric` section determines which of the available evaluation metrics is goi Currently, the following metrics are supported: [float] +[[k-precision]] ==== Precision at K (P@k) This metric measures the number of relevant results in the top k search results. Its a form of the well known https://en.wikipedia.org/wiki/Information_retrieval#Precision[Precision] metric that only looks at the top k documents. It is the fraction of relevant documents in those first k diff --git a/docs/reference/settings/ssl-settings.asciidoc b/docs/reference/settings/ssl-settings.asciidoc index a4422b8fb2d3c..f392d0c2fb816 100644 --- a/docs/reference/settings/ssl-settings.asciidoc +++ b/docs/reference/settings/ssl-settings.asciidoc @@ -39,7 +39,6 @@ endif::verifies[] Supported cipher suites can be found in Oracle's http://docs.oracle.com/javase/8/docs/technotes/guides/security/SunProviders.html[ Java Cryptography Architecture documentation]. Defaults to ``. - ===== {component} TLS/SSL Key and Trusted Certificate Settings The following settings are used to specify a private key, certificate, and the diff --git a/docs/reference/setup/install/deb.asciidoc b/docs/reference/setup/install/deb.asciidoc index 192737581bd57..529c6b12ce520 100644 --- a/docs/reference/setup/install/deb.asciidoc +++ b/docs/reference/setup/install/deb.asciidoc @@ -173,6 +173,7 @@ include::xpack-indices.asciidoc[] endif::include-xpack[] +[[deb-sysv-init-vs-systemd]] ==== SysV `init` vs `systemd` include::init-systemd.asciidoc[] diff --git a/docs/reference/setup/install/docker.asciidoc b/docs/reference/setup/install/docker.asciidoc index 76112f220dc3a..172c7c1f17c2f 100644 --- a/docs/reference/setup/install/docker.asciidoc +++ b/docs/reference/setup/install/docker.asciidoc @@ -287,6 +287,7 @@ comfortable with them adding the `--batch` flag to the plugin install command. See {plugins}/_other_command_line_parameters.html[Plugin Management documentation] for more details. +[[override-image-default]] ===== D. Override the image's default https://docs.docker.com/engine/reference/run/#cmd-default-command-or-options[CMD] Options can be passed as command-line options to the {es} process by diff --git a/docs/reference/setup/install/rpm.asciidoc b/docs/reference/setup/install/rpm.asciidoc index c6f5647d08e26..041784bb96862 100644 --- a/docs/reference/setup/install/rpm.asciidoc +++ b/docs/reference/setup/install/rpm.asciidoc @@ -160,6 +160,7 @@ include::xpack-indices.asciidoc[] endif::include-xpack[] +[[rpm-sysv-init-vs-systemd]] ==== SysV `init` vs `systemd` include::init-systemd.asciidoc[] diff --git a/docs/reference/setup/starting.asciidoc b/docs/reference/setup/starting.asciidoc index 4474c247361db..c2ab6b1404aa2 100644 --- a/docs/reference/setup/starting.asciidoc +++ b/docs/reference/setup/starting.asciidoc @@ -34,9 +34,11 @@ include::install/zip-windows-start.asciidoc[] include::install/init-systemd.asciidoc[] [float] +[[start-es-deb-init]] include::install/deb-init.asciidoc[] [float] +[[start-es-deb-systemd]] include::install/systemd.asciidoc[] [float] @@ -66,7 +68,9 @@ include::install/msi-windows-start.asciidoc[] include::install/init-systemd.asciidoc[] [float] +[[start-es-rpm-init]] include::install/rpm-init.asciidoc[] [float] +[[start-es-rpm-systemd]] include::install/systemd.asciidoc[] diff --git a/docs/reference/sql/endpoints/jdbc.asciidoc b/docs/reference/sql/endpoints/jdbc.asciidoc index 9014d3f5719b5..33f130a891896 100644 --- a/docs/reference/sql/endpoints/jdbc.asciidoc +++ b/docs/reference/sql/endpoints/jdbc.asciidoc @@ -156,6 +156,7 @@ Opens up a {es-sql} connection to `server` on port `3456`, setting the JDBC conn One can use JDBC through the official `java.sql` and `javax.sql` packages: +[[java-sql]] ==== `java.sql` The former through `java.sql.Driver` and `DriverManager`: @@ -168,6 +169,7 @@ HTTP traffic. The port is by default 9200. <2> Properties for connecting to Elasticsearch. An empty `Properties` instance is fine for unsecured Elasticsearch. +[[javax-sql]] ==== `javax.sql` Accessible through the `javax.sql.DataSource` API: diff --git a/docs/reference/sql/limitations.asciidoc b/docs/reference/sql/limitations.asciidoc index e2a538cd08571..8f7868c892e52 100644 --- a/docs/reference/sql/limitations.asciidoc +++ b/docs/reference/sql/limitations.asciidoc @@ -4,6 +4,7 @@ == SQL Limitations [float] +[[sys-columns-describe-table-nested-fields]] === Nested fields in `SYS COLUMNS` and `DESCRIBE TABLE` {es} has a special type of relationship fields called `nested` fields. In {es-sql} they can be used by referencing their inner @@ -51,6 +52,7 @@ This is because of the way nested queries work in {es}: the root nested field wi pagination taking place on the **root nested document and not on its inner hits**. [float] +[[normalized-keyword-fields]] === Normalized `keyword` fields `keyword` fields in {es} can be normalized by defining a `normalizer`. Such fields are not supported in {es-sql}. @@ -108,6 +110,7 @@ But, if the sub-select would include a `GROUP BY` or `HAVING` or the enclosing ` FROM (SELECT ...) WHERE [simple_condition]`, this is currently **un-supported**. [float] +[[first-last-agg-functions-having-clause]] === Using <>/<> aggregation functions in `HAVING` clause Using `FIRST` and `LAST` in the `HAVING` clause is not supported. The same applies to @@ -115,6 +118,7 @@ Using `FIRST` and `LAST` in the `HAVING` clause is not supported. The same appli is of type <> as they are internally translated to `FIRST` and `LAST`. [float] +[[group-by-time]] === Using TIME data type in GROUP BY or <> Using `TIME` data type as a grouping key is currently not supported. For example: diff --git a/docs/reference/sql/security.asciidoc b/docs/reference/sql/security.asciidoc index a317355866b8b..ad946c33e2d98 100644 --- a/docs/reference/sql/security.asciidoc +++ b/docs/reference/sql/security.asciidoc @@ -7,6 +7,7 @@ In such a scenario, {es-sql} supports both security at the transport layer (by encrypting the communication between the consumer and the server) and authentication (for the access layer). [float] +[[ssl-tls-config]] ==== SSL/TLS configuration In case of an encrypted transport, the SSL/TLS support needs to be enabled in {es-sql} to properly establish communication with {es}. This is done by setting the `ssl` property to `true` or by using the `https` prefix in the URL. +