mirror of
https://github.com/elastic/elasticsearch.git
synced 2025-06-28 17:34:17 -04:00
Other tokenizers like the standard tokenizer allow overriding the default maximum token length of 255 using the `"max_token_length` parameter. This change enables using this parameter also with the whitespace tokenizer. The range that is currently allowed is from 0 to StandardTokenizer.MAX_TOKEN_LENGTH_LIMIT, which is 1024 * 1024 = 1048576 characters. Closes #26643 |
||
---|---|---|
.. | ||
classic-tokenizer.asciidoc | ||
edgengram-tokenizer.asciidoc | ||
keyword-tokenizer.asciidoc | ||
letter-tokenizer.asciidoc | ||
lowercase-tokenizer.asciidoc | ||
ngram-tokenizer.asciidoc | ||
pathhierarchy-tokenizer.asciidoc | ||
pattern-tokenizer.asciidoc | ||
simplepattern-tokenizer.asciidoc | ||
simplepatternsplit-tokenizer.asciidoc | ||
standard-tokenizer.asciidoc | ||
thai-tokenizer.asciidoc | ||
uaxurlemail-tokenizer.asciidoc | ||
whitespace-tokenizer.asciidoc |