| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677 | [[analysis]]= Analysis[partintro]--The index analysis module acts as a configurable registry of Analyzersthat can be used in order to both break indexed (analyzed) fields when adocument is indexed and process query strings. It maps to the Lucene`Analyzer`.Analyzers are composed of a single <<analysis-tokenizers,Tokenizer>> and zero or more <<analysis-tokenfilters,TokenFilters>>. The tokenizer may be preceded by one or more <<analysis-charfilters,CharFilters>>. Theanalysis module allows one to register `TokenFilters`, `Tokenizers` and`Analyzers` under logical names that can then be referenced either inmapping definitions or in certain APIs. The Analysis moduleautomatically registers (*if not explicitly defined*) built inanalyzers, token filters, and tokenizers.Here is a sample configuration:[source,js]--------------------------------------------------index :    analysis :        analyzer :             standard :                 type : standard                stopwords : [stop1, stop2]            myAnalyzer1 :                type : standard                stopwords : [stop1, stop2, stop3]                max_token_length : 500            # configure a custom analyzer which is             # exactly like the default standard analyzer            myAnalyzer2 :                tokenizer : standard                filter : [standard, lowercase, stop]        tokenizer :            myTokenizer1 :                type : standard                max_token_length : 900            myTokenizer2 :                type : keyword                buffer_size : 512        filter :            myTokenFilter1 :                type : stop                stopwords : [stop1, stop2, stop3, stop4]            myTokenFilter2 :                type : length                min : 0                max : 2000--------------------------------------------------[float][[backwards-compatibility]]=== Backwards compatibilityAll analyzers, tokenizers, and token filters can be configured with a`version` parameter to control which Lucene version behavior they shoulduse. Possible values are: `3.0` - `3.6`, `4.0` - `4.3` (the highestversion number is the default option).--include::analysis/analyzers.asciidoc[]include::analysis/tokenizers.asciidoc[]include::analysis/tokenfilters.asciidoc[]include::analysis/charfilters.asciidoc[]include::analysis/icu-plugin.asciidoc[]
 |