elasticsearch/docs/reference/analysis
James Rodewig 2fe9ba53ec
[DOCS] Note limitations of max_gram parm in edge_ngram tokenizer for index analyzers (#49007)
The `edge_ngram` tokenizer limits tokens to the `max_gram` character
length. Autocomplete searches for terms longer than this limit return
no results.

To prevent this, you can use the `truncate` token filter to truncate
tokens to the `max_gram` character length. However, this could return irrelevant results.

This commit adds some advisory text to make users aware of this limitation and outline the tradeoffs for each approach.

Closes #48956.
2019-11-13 14:27:10 -05:00
..
analyzers [DOCS] Remove unneeded spaces from custom analyzer snippet (#47332) 2019-10-15 15:52:52 -04:00
charfilters Remove _type from SearchHit (#46942) 2019-09-23 19:14:54 +01:00
tokenfilters [DOCS] Reformat compound word token filters (#49006) 2019-11-13 09:35:00 -05:00
tokenizers [DOCS] Note limitations of max_gram parm in edge_ngram tokenizer for index analyzers (#49007) 2019-11-13 14:27:10 -05:00
analyzers.asciidoc [DOCS] Sort analyzers, tokenizers, and token filters alphabetically (#48068) 2019-10-15 15:46:50 -04:00
anatomy.asciidoc Correction of the names of numirals (#21531) 2016-11-25 14:30:49 +01:00
charfilters.asciidoc Hindu-Arabico-Latino Numerals (#22476) 2017-01-10 15:24:56 +01:00
normalizers.asciidoc [DOCS] [2 of 5] Change // CONSOLE comments to [source,console] (#46353) 2019-09-09 13:13:41 -04:00
testing.asciidoc [DOCS] [2 of 5] Change // CONSOLE comments to [source,console] (#46353) 2019-09-09 13:13:41 -04:00
tokenfilters.asciidoc [DOCS] Reformat compound word token filters (#49006) 2019-11-13 09:35:00 -05:00
tokenizers.asciidoc [DOCS] Sort analyzers, tokenizers, and token filters alphabetically (#48068) 2019-10-15 15:46:50 -04:00