| 123456789101112131415161718 | [[analysis-standard-tokenizer]]=== Standard TokenizerA tokenizer of type `standard` providing grammar based tokenizer that isa good tokenizer for most European language documents. The tokenizerimplements the Unicode Text Segmentation algorithm, as specified inhttp://unicode.org/reports/tr29/[Unicode Standard Annex #29].The following are settings that can be set for a `standard` tokenizertype:[cols="<,<",options="header",]|=======================================================================|Setting |Description|`max_token_length` |The maximum token length. If a token is seen thatexceeds this length then it is discarded. Defaults to `255`.|=======================================================================
 |