| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960 | [[analysis-keyword-tokenizer]]=== Keyword TokenizerThe `keyword` tokenizer  is a ``noop'' tokenizer that accepts whatever text itis given and outputs the exact same text as a single term.  It can be combinedwith token filters to normalise output, e.g. lower-casing email addresses.[float]=== Example output[source,js]---------------------------POST _analyze{  "tokenizer": "keyword",  "text": "New York"}---------------------------// CONSOLE/////////////////////[source,js]----------------------------{  "tokens": [    {      "token": "New York",      "start_offset": 0,      "end_offset": 8,      "type": "word",      "position": 0    }  ]}----------------------------// TESTRESPONSE/////////////////////The above sentence would produce the following term:[source,text]---------------------------[ New York ]---------------------------[float]=== ConfigurationThe `keyword` tokenizer accepts the following parameters:[horizontal]`buffer_size`::    The number of characters read into the term buffer in a single pass.    Defaults to `256`.  The term buffer will grow by this size until all the    text has been consumed.  It is advisable not to change this setting.
 |