Browse Source

Update experimental labels in the docs (#25727)

Relates https://github.com/elastic/elasticsearch/issues/19798

Removed experimental label from:
* Painless
* Diversified Sampler Agg
* Sampler Agg
* Significant Terms Agg
* Terms Agg document count error and execution_hint
* Cardinality Agg precision_threshold
* Pipeline Aggregations
* index.shard.check_on_startup
* index.store.type (added warning)
* Preloading data into the file system cache
* foreach ingest processor
* Field caps API
* Profile API

Added experimental label to:
* Moving Average Agg Prediction


Changed experimental to beta for:
* Adjacency matrix agg
* Normalizers
* Tasks API
* Index sorting

Labelled experimental in Lucene:
* ICU plugin custom rules file
* Flatten graph token filter
* Synonym graph token filter
* Word delimiter graph token filter
* Simple pattern tokenizer
* Simple pattern split tokenizer

Replaced experimental label with warning that details may change in the future:
* Analysis explain output format
* Segments verbose output format
* Percentile Agg compression and HDR Histogram
* Percentile Rank Agg HDR Histogram
Clinton Gormley 8 years ago
parent
commit
ff4a2519f2
43 changed files with 22 additions and 78 deletions
  1. 0 2
      docs/painless/painless-debugging.asciidoc
  2. 0 2
      docs/painless/painless-getting-started.asciidoc
  3. 0 2
      docs/painless/painless-syntax.asciidoc
  4. 1 1
      docs/plugins/analysis-icu.asciidoc
  5. 1 1
      docs/reference/aggregations/bucket/adjacency-matrix-aggregation.asciidoc
  6. 0 2
      docs/reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc
  7. 0 2
      docs/reference/aggregations/bucket/sampler-aggregation.asciidoc
  8. 0 2
      docs/reference/aggregations/bucket/significantterms-aggregation.asciidoc
  9. 1 5
      docs/reference/aggregations/bucket/terms-aggregation.asciidoc
  10. 0 2
      docs/reference/aggregations/metrics/cardinality-aggregation.asciidoc
  11. 1 3
      docs/reference/aggregations/metrics/percentile-aggregation.asciidoc
  12. 1 1
      docs/reference/aggregations/metrics/percentile-rank-aggregation.asciidoc
  13. 0 2
      docs/reference/aggregations/pipeline.asciidoc
  14. 0 2
      docs/reference/aggregations/pipeline/avg-bucket-aggregation.asciidoc
  15. 0 2
      docs/reference/aggregations/pipeline/bucket-script-aggregation.asciidoc
  16. 0 2
      docs/reference/aggregations/pipeline/bucket-selector-aggregation.asciidoc
  17. 0 2
      docs/reference/aggregations/pipeline/cumulative-sum-aggregation.asciidoc
  18. 0 2
      docs/reference/aggregations/pipeline/derivative-aggregation.asciidoc
  19. 0 2
      docs/reference/aggregations/pipeline/extended-stats-bucket-aggregation.asciidoc
  20. 0 2
      docs/reference/aggregations/pipeline/max-bucket-aggregation.asciidoc
  21. 0 2
      docs/reference/aggregations/pipeline/min-bucket-aggregation.asciidoc
  22. 2 2
      docs/reference/aggregations/pipeline/movavg-aggregation.asciidoc
  23. 0 2
      docs/reference/aggregations/pipeline/percentiles-bucket-aggregation.asciidoc
  24. 0 2
      docs/reference/aggregations/pipeline/serial-diff-aggregation.asciidoc
  25. 0 2
      docs/reference/aggregations/pipeline/stats-bucket-aggregation.asciidoc
  26. 0 2
      docs/reference/aggregations/pipeline/sum-bucket-aggregation.asciidoc
  27. 1 1
      docs/reference/analysis/normalizers.asciidoc
  28. 1 1
      docs/reference/analysis/tokenfilters/flatten-graph-tokenfilter.asciidoc
  29. 1 1
      docs/reference/analysis/tokenfilters/synonym-graph-tokenfilter.asciidoc
  30. 1 1
      docs/reference/analysis/tokenfilters/word-delimiter-graph-tokenfilter.asciidoc
  31. 1 1
      docs/reference/analysis/tokenizers/simplepattern-tokenizer.asciidoc
  32. 1 1
      docs/reference/analysis/tokenizers/simplepatternsplit-tokenizer.asciidoc
  33. 1 1
      docs/reference/cluster/tasks.asciidoc
  34. 2 2
      docs/reference/index-modules.asciidoc
  35. 1 1
      docs/reference/index-modules/index-sorting.asciidoc
  36. 2 2
      docs/reference/index-modules/store.asciidoc
  37. 1 1
      docs/reference/indices/analyze.asciidoc
  38. 1 1
      docs/reference/indices/segments.asciidoc
  39. 0 5
      docs/reference/ingest/ingest-node.asciidoc
  40. 0 1
      docs/reference/mapping/types/keyword.asciidoc
  41. 0 2
      docs/reference/modules/scripting/painless.asciidoc
  42. 0 2
      docs/reference/search/field-caps.asciidoc
  43. 1 1
      docs/reference/search/profile.asciidoc

+ 0 - 2
docs/painless/painless-debugging.asciidoc

@@ -1,8 +1,6 @@
 [[painless-debugging]]
 === Painless Debugging
 
-experimental[The Painless scripting language is new and is still marked as experimental. The syntax or API may be changed in the future in non-backwards compatible ways if required.]
-
 ==== Debug.Explain
 
 Painless doesn't have a

+ 0 - 2
docs/painless/painless-getting-started.asciidoc

@@ -1,8 +1,6 @@
 [[painless-getting-started]]
 == Getting Started with Painless
 
-experimental[The Painless scripting language is new and is still marked as experimental. The syntax or API may be changed in the future in non-backwards compatible ways if required.]
-
 include::painless-description.asciidoc[]
 
 [[painless-examples]]

+ 0 - 2
docs/painless/painless-syntax.asciidoc

@@ -1,8 +1,6 @@
 [[painless-syntax]]
 === Painless Syntax
 
-experimental[The Painless scripting language is new and is still marked as experimental. The syntax or API may be changed in the future in non-backwards compatible ways if required.]
-
 [float]
 [[control-flow]]
 ==== Control flow

+ 1 - 1
docs/plugins/analysis-icu.asciidoc

@@ -113,7 +113,7 @@ PUT icu_sample
 
 ===== Rules customization
 
-experimental[]
+experimental[This functionality is marked as experimental in Lucene]
 
 You can customize the `icu-tokenizer` behavior by specifying per-script rule files, see the
 http://userguide.icu-project.org/boundaryanalysis#TOC-RBBI-Rules[RBBI rules syntax reference]

+ 1 - 1
docs/reference/aggregations/bucket/adjacency-matrix-aggregation.asciidoc

@@ -6,7 +6,7 @@ The request provides a collection of named filter expressions, similar to the `f
 request. 
 Each bucket in the response represents a non-empty cell in the matrix of intersecting filters.
 
-experimental[The `adjacency_matrix` aggregation is a new feature and we may evolve its design as we get feedback on its use.  As a result, the API for this feature may change in non-backwards compatible ways]
+beta[The `adjacency_matrix` aggregation is a new feature and we may evolve its design as we get feedback on its use.  As a result, the API for this feature may change in non-backwards compatible ways]
 
 
 Given filters named `A`, `B` and `C` the response would return buckets with the following names:

+ 0 - 2
docs/reference/aggregations/bucket/diversified-sampler-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-bucket-diversified-sampler-aggregation]]
 === Diversified Sampler Aggregation
 
-experimental[]
-
 Like the `sampler` aggregation this is a filtering aggregation used to limit any sub aggregations' processing to a sample of the top-scoring documents.
 The `diversified_sampler` aggregation adds the ability to limit the number of matches that share a common value such as an "author".
 

+ 0 - 2
docs/reference/aggregations/bucket/sampler-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-bucket-sampler-aggregation]]
 === Sampler Aggregation
 
-experimental[]
-
 A filtering aggregation used to limit any sub aggregations' processing to a sample of the top-scoring documents.
 
 .Example use cases:

+ 0 - 2
docs/reference/aggregations/bucket/significantterms-aggregation.asciidoc

@@ -3,8 +3,6 @@
 
 An aggregation that returns interesting or unusual occurrences of terms in a set.
 
-experimental[The `significant_terms` aggregation can be very heavy when run on large indices.  Work is in progress to provide more lightweight sampling techniques.  As a result, the API for this feature may change in non-backwards compatible ways]
-
 .Example use cases:
 * Suggesting "H5N1" when users search for "bird flu" in text
 * Identifying the merchant that is the "common point of compromise" from the transaction history of credit card owners reporting loss

+ 1 - 5
docs/reference/aggregations/bucket/terms-aggregation.asciidoc

@@ -197,8 +197,6 @@ could have the 4th highest document count.
 
 ==== Per bucket document count error
 
-experimental[]
-
 The second error value can be enabled by setting the `show_term_doc_count_error` parameter to true. This shows an error value
 for each term returned by the aggregation which represents the 'worst case' error in the document count and can be useful when
 deciding on a value for the `shard_size` parameter. This is calculated by summing the document counts for the last term returned
@@ -728,8 +726,6 @@ collection mode need to replay the query on the second pass but only for the doc
 [[search-aggregations-bucket-terms-aggregation-execution-hint]]
 ==== Execution hint
 
-experimental[The automated execution optimization is experimental, so this parameter is provided temporarily as a way to override the default behaviour]
-
 There are different mechanisms by which terms aggregations can be executed:
 
  - by using field values directly in order to aggregate data per-bucket (`map`)
@@ -767,7 +763,7 @@ in inner aggregations.
 }
 --------------------------------------------------
 
-<1> experimental[] the possible values are `map`, `global_ordinals`, `global_ordinals_hash` and `global_ordinals_low_cardinality`
+<1> The possible values are `map`, `global_ordinals`, `global_ordinals_hash` and `global_ordinals_low_cardinality`
 
 Please note that Elasticsearch will ignore this execution hint if it is not applicable and that there is no backward compatibility guarantee on these hints.
 

+ 0 - 2
docs/reference/aggregations/metrics/cardinality-aggregation.asciidoc

@@ -43,8 +43,6 @@ Response:
 
 This aggregation also supports the `precision_threshold` option:
 
-experimental[The `precision_threshold` option is specific to the current internal implementation of the `cardinality` agg, which may change in the future]
-
 [source,js]
 --------------------------------------------------
 POST /sales/_search?size=0

+ 1 - 3
docs/reference/aggregations/metrics/percentile-aggregation.asciidoc

@@ -247,8 +247,6 @@ it. It would not be the case on more skewed distributions.
 [[search-aggregations-metrics-percentile-aggregation-compression]]
 ==== Compression
 
-experimental[The `compression` parameter is specific to the current internal implementation of percentiles, and may change in the future]
-
 Approximate algorithms must balance memory utilization with estimation accuracy.
 This balance can be controlled using a `compression` parameter:
 
@@ -287,7 +285,7 @@ the TDigest will use less memory.
 
 ==== HDR Histogram
 
-experimental[]
+NOTE: This setting exposes the internal implementation of HDR Histogram and the syntax may change in the future.
 
 https://github.com/HdrHistogram/HdrHistogram[HDR Histogram] (High Dynamic Range Histogram) is an alternative implementation
 that can be useful when calculating percentiles for latency measurements as it can be faster than the t-digest implementation

+ 1 - 1
docs/reference/aggregations/metrics/percentile-rank-aggregation.asciidoc

@@ -159,7 +159,7 @@ This will interpret the `script` parameter as an `inline` script with the `painl
 
 ==== HDR Histogram
 
-experimental[]
+NOTE: This setting exposes the internal implementation of HDR Histogram and the syntax may change in the future.
 
 https://github.com/HdrHistogram/HdrHistogram[HDR Histogram] (High Dynamic Range Histogram) is an alternative implementation
 that can be useful when calculating percentile ranks for latency measurements as it can be faster than the t-digest implementation

+ 0 - 2
docs/reference/aggregations/pipeline.asciidoc

@@ -2,8 +2,6 @@
 
 == Pipeline Aggregations
 
-experimental[]
-
 Pipeline aggregations work on the outputs produced from other aggregations rather than from document sets, adding
 information to the output tree. There are many different types of pipeline aggregation, each computing different information from
 other aggregations, but these types can be broken down into two families:

+ 0 - 2
docs/reference/aggregations/pipeline/avg-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-avg-bucket-aggregation]]
 === Avg Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which calculates the (mean) average value of a specified metric in a sibling aggregation.
 The specified metric must be numeric and the sibling aggregation must be a multi-bucket aggregation.
 

+ 0 - 2
docs/reference/aggregations/pipeline/bucket-script-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-bucket-script-aggregation]]
 === Bucket Script Aggregation
 
-experimental[]
-
 A parent pipeline aggregation which executes a script which can perform per bucket computations on specified metrics
 in the parent multi-bucket aggregation. The specified metric must be numeric and the script must return a numeric value.
 

+ 0 - 2
docs/reference/aggregations/pipeline/bucket-selector-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-bucket-selector-aggregation]]
 === Bucket Selector Aggregation
 
-experimental[]
-
 A parent pipeline aggregation which executes a script which determines whether the current bucket will be retained
 in the parent multi-bucket aggregation. The specified metric must be numeric and the script must return a boolean value.
 If the script language is `expression` then a numeric return value is permitted. In this case 0.0 will be evaluated as `false`

+ 0 - 2
docs/reference/aggregations/pipeline/cumulative-sum-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-cumulative-sum-aggregation]]
 === Cumulative Sum Aggregation
 
-experimental[]
-
 A parent pipeline aggregation which calculates the cumulative sum of a specified metric in a parent histogram (or date_histogram)
 aggregation. The specified metric must be numeric and the enclosing histogram must have `min_doc_count` set to `0` (default
 for `histogram` aggregations).

+ 0 - 2
docs/reference/aggregations/pipeline/derivative-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-derivative-aggregation]]
 === Derivative Aggregation
 
-experimental[]
-
 A parent pipeline aggregation which calculates the derivative of a specified metric in a parent histogram (or date_histogram)
 aggregation. The specified metric must be numeric and the enclosing histogram must have `min_doc_count` set to `0` (default
 for `histogram` aggregations).

+ 0 - 2
docs/reference/aggregations/pipeline/extended-stats-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-extended-stats-bucket-aggregation]]
 === Extended Stats Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which calculates a variety of stats across all bucket of a specified metric in a sibling aggregation.
 The specified metric must be numeric and the sibling aggregation must be a multi-bucket aggregation.
 

+ 0 - 2
docs/reference/aggregations/pipeline/max-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-max-bucket-aggregation]]
 === Max Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which identifies the bucket(s) with the maximum value of a specified metric in a sibling aggregation
 and outputs both the value and the key(s) of the bucket(s). The specified metric must be numeric and the sibling aggregation must
 be a multi-bucket aggregation.

+ 0 - 2
docs/reference/aggregations/pipeline/min-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-min-bucket-aggregation]]
 === Min Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which identifies the bucket(s) with the minimum value of a specified metric in a sibling aggregation
 and outputs both the value and the key(s) of the bucket(s). The specified metric must be numeric and the sibling aggregation must
 be a multi-bucket aggregation.

+ 2 - 2
docs/reference/aggregations/pipeline/movavg-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-movavg-aggregation]]
 === Moving Average Aggregation
 
-experimental[]
-
 Given an ordered series of data, the Moving Average aggregation will slide a window across the data and emit the average
 value of that window.  For example, given the data `[1, 2, 3, 4, 5, 6, 7, 8, 9, 10]`, we can calculate a simple moving
 average with windows size of `5` as follows:
@@ -513,6 +511,8 @@ POST /_search
 
 ==== Prediction
 
+experimental[]
+
 All the moving average model support a "prediction" mode, which will attempt to extrapolate into the future given the
 current smoothed, moving average.  Depending on the model and parameter, these predictions may or may not be accurate.
 

+ 0 - 2
docs/reference/aggregations/pipeline/percentiles-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-percentiles-bucket-aggregation]]
 === Percentiles Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which calculates percentiles across all bucket of a specified metric in a sibling aggregation.
 The specified metric must be numeric and the sibling aggregation must be a multi-bucket aggregation.
 

+ 0 - 2
docs/reference/aggregations/pipeline/serial-diff-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-serialdiff-aggregation]]
 === Serial Differencing Aggregation
 
-experimental[]
-
 Serial differencing is a technique where values in a time series are subtracted from itself at
 different time lags or periods. For example, the datapoint f(x) = f(x~t~) - f(x~t-n~), where n is the period being used.
 

+ 0 - 2
docs/reference/aggregations/pipeline/stats-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-stats-bucket-aggregation]]
 === Stats Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which calculates a variety of stats across all bucket of a specified metric in a sibling aggregation.
 The specified metric must be numeric and the sibling aggregation must be a multi-bucket aggregation.
 

+ 0 - 2
docs/reference/aggregations/pipeline/sum-bucket-aggregation.asciidoc

@@ -1,8 +1,6 @@
 [[search-aggregations-pipeline-sum-bucket-aggregation]]
 === Sum Bucket Aggregation
 
-experimental[]
-
 A sibling pipeline aggregation which calculates the sum across all bucket of a specified metric in a sibling aggregation.
 The specified metric must be numeric and the sibling aggregation must be a multi-bucket aggregation.
 

+ 1 - 1
docs/reference/analysis/normalizers.asciidoc

@@ -1,7 +1,7 @@
 [[analysis-normalizers]]
 == Normalizers
 
-experimental[]
+beta[]
 
 Normalizers are similar to analyzers except that they may only emit a single
 token. As a consequence, they do not have a tokenizer and only accept a subset

+ 1 - 1
docs/reference/analysis/tokenfilters/flatten-graph-tokenfilter.asciidoc

@@ -1,7 +1,7 @@
 [[analysis-flatten-graph-tokenfilter]]
 === Flatten Graph Token Filter
 
-experimental[]
+experimental[This functionality is marked as experimental in Lucene]
 
 The `flatten_graph` token filter accepts an arbitrary graph token
 stream, such as that produced by

+ 1 - 1
docs/reference/analysis/tokenfilters/synonym-graph-tokenfilter.asciidoc

@@ -1,7 +1,7 @@
 [[analysis-synonym-graph-tokenfilter]]
 === Synonym Graph Token Filter
 
-experimental[]
+experimental[This functionality is marked as experimental in Lucene]
 
 The `synonym_graph` token filter allows to easily handle synonyms,
 including multi-word synonyms correctly during the analysis process.

+ 1 - 1
docs/reference/analysis/tokenfilters/word-delimiter-graph-tokenfilter.asciidoc

@@ -1,7 +1,7 @@
 [[analysis-word-delimiter-graph-tokenfilter]]
 === Word Delimiter Graph Token Filter
 
-experimental[]
+experimental[This functionality is marked as experimental in Lucene]
 
 Named `word_delimiter_graph`, it splits words into subwords and performs
 optional transformations on subword groups. Words are split into

+ 1 - 1
docs/reference/analysis/tokenizers/simplepattern-tokenizer.asciidoc

@@ -1,7 +1,7 @@
 [[analysis-simplepattern-tokenizer]]
 === Simple Pattern Tokenizer
 
-experimental[]
+experimental[This functionality is marked as experimental in Lucene]
 
 The `simple_pattern` tokenizer uses a regular expression to capture matching
 text as terms. The set of regular expression features it supports is more

+ 1 - 1
docs/reference/analysis/tokenizers/simplepatternsplit-tokenizer.asciidoc

@@ -1,7 +1,7 @@
 [[analysis-simplepatternsplit-tokenizer]]
 === Simple Pattern Split Tokenizer
 
-experimental[]
+experimental[This functionality is marked as experimental in Lucene]
 
 The `simple_pattern_split` tokenizer uses a regular expression to split the
 input into terms at pattern matches. The set of regular expression features it

+ 1 - 1
docs/reference/cluster/tasks.asciidoc

@@ -1,7 +1,7 @@
 [[tasks]]
 == Task Management API
 
-experimental[The Task Management API is new and should still be considered experimental.  The API may change in ways that are not backwards compatible]
+beta[The Task Management API is new and should still be considered a beta feature.  The API may change in ways that are not backwards compatible]
 
 [float]
 === Current Tasks Information

+ 2 - 2
docs/reference/index-modules.asciidoc

@@ -47,7 +47,7 @@ specific index module:
 `index.shard.check_on_startup`::
 +
 --
-experimental[] Whether or not shards should be checked for corruption before opening. When
+Whether or not shards should be checked for corruption before opening. When
 corruption is detected, it will prevent the shard from being opened. Accepts:
 
 `false`::
@@ -69,7 +69,7 @@ corruption is detected, it will prevent the shard from being opened. Accepts:
     as corrupted will be automatically removed. This option *may result in data loss*.
     Use with extreme caution!
 
-Checking shards may take a lot of time on large indices.
+WARNING: Expert only. Checking shards may take a lot of time on large indices.
 --
 
 [[index-codec]] `index.codec`::

+ 1 - 1
docs/reference/index-modules/index-sorting.asciidoc

@@ -1,7 +1,7 @@
 [[index-modules-index-sorting]]
 == Index Sorting
 
-experimental[]
+beta[]
 
 When creating a new index in elasticsearch it is possible to configure how the Segments
 inside each Shard will be sorted. By default Lucene does not apply any sort.

+ 2 - 2
docs/reference/index-modules/store.asciidoc

@@ -32,7 +32,7 @@ PUT /my_index
 }
 ---------------------------------
 
-experimental[This is an expert-only setting and may be removed in the future]
+WARNING: This is an expert-only setting and may be removed in the future.
 
 The following sections lists all the different storage types supported.
 
@@ -73,7 +73,7 @@ compatibility.
 
 === Pre-loading data into the file system cache
 
-experimental[This is an expert-only setting and may be removed in the future]
+NOTE: This is an expert setting, the details of which may change in the future.
 
 By default, elasticsearch completely relies on the operating system file system
 cache for caching I/O operations. It is possible to set `index.store.preload`

+ 1 - 1
docs/reference/indices/analyze.asciidoc

@@ -144,7 +144,7 @@ GET _analyze
 If you want to get more advanced details, set `explain` to `true` (defaults to `false`). It will output all token attributes for each token.
 You can filter token attributes you want to output by setting `attributes` option.
 
-experimental[The format of the additional detail information is experimental and can change at any time]
+NOTE: The format of the additional detail information is labelled as experimental in Lucene and it may change in the future.
 
 [source,js]
 --------------------------------------------------

+ 1 - 1
docs/reference/indices/segments.asciidoc

@@ -79,7 +79,7 @@ compound::   Whether the segment is stored in a compound file. When true, this
 
 To add additional information that can be used for debugging, use the `verbose` flag.
 
-experimental[The format of the additional verbose information is experimental and can change at any time]
+NOTE: The format of the additional detail information is labelled as experimental in Lucene and it may change in the future.
 
 [source,js]
 --------------------------------------------------

+ 0 - 5
docs/reference/ingest/ingest-node.asciidoc

@@ -1010,11 +1010,6 @@ to the requester.
 [[foreach-processor]]
 === Foreach Processor
 
-experimental[This processor may change or be replaced by something else that provides similar functionality. This
-processor executes in its own context, which makes it different compared to all other processors and for features like
-verbose simulation the subprocessor isn't visible. The reason we still expose this processor, is that it is the only
-processor that can operate on an array]
-
 Processes elements in an array of unknown length.
 
 All processors can operate on elements inside an array, but if all elements of an array need to

+ 0 - 1
docs/reference/mapping/types/keyword.asciidoc

@@ -98,7 +98,6 @@ The following parameters are accepted by `keyword` fields:
 
 <<normalizer,`normalizer`>>::
 
-    experimental[]
     How to pre-process the keyword prior to indexing. Defaults to `null`,
     meaning the keyword is kept as-is.
 

+ 0 - 2
docs/reference/modules/scripting/painless.asciidoc

@@ -1,8 +1,6 @@
 [[modules-scripting-painless]]
 === Painless Scripting Language
 
-experimental[The Painless scripting language is new and is still marked as experimental. The syntax or API may be changed in the future in non-backwards compatible ways if required.]
-
 include::../../../painless/painless-description.asciidoc[]
 
 Ready to start scripting with Painless? See {painless}/painless-getting-started.html[Getting Started with Painless] in the guide to the

+ 0 - 2
docs/reference/search/field-caps.asciidoc

@@ -1,8 +1,6 @@
 [[search-field-caps]]
 == Field Capabilities API
 
-experimental[]
-
 The field capabilities API allows to retrieve the capabilities of fields among multiple indices.
 
 The field capabilities api by default executes on all indices:

+ 1 - 1
docs/reference/search/profile.asciidoc

@@ -1,7 +1,7 @@
 [[search-profile]]
 == Profile API
 
-experimental[]
+WARNING:  The Profile API is a debugging tool and adds signficant overhead to search execution.
 
 The Profile API provides detailed timing information about the execution of individual components
 in a search request.  It gives the user insight into how search requests are executed at a low level so that