elasticsearch/docs/reference/elasticsearch-plugins/analysis-kuromoji-completion.md
Colleen McGinnis b7e3a1e14b
[docs] Migrate docs from AsciiDoc to Markdown (#123507)
* delete asciidoc files

* add migrated files

* fix errors

* Disable docs tests

* Clarify release notes page titles

* Revert "Clarify release notes page titles"

This reverts commit 8be688648d.

* Comment out edternal URI images

* Clean up query languages landing pages, link to conceptual docs

* Add .md to url

* Fixes inference processor nesting.

---------

Co-authored-by: Liam Thompson <32779855+leemthompo@users.noreply.github.com>
Co-authored-by: Liam Thompson <leemthompo@gmail.com>
Co-authored-by: Martijn Laarman <Mpdreamz@gmail.com>
Co-authored-by: István Zoltán Szabó <szabosteve@gmail.com>
2025-02-27 17:56:14 +01:00

861 B

mapped_pages
https://www.elastic.co/guide/en/elasticsearch/plugins/current/analysis-kuromoji-completion.html

kuromoji_completion token filter [analysis-kuromoji-completion]

The kuromoji_completion token filter adds Japanese romanized tokens to the term attributes along with the original tokens (surface forms).

GET _analyze
{
  "analyzer": "kuromoji_completion",
  "text": "寿司" <1>
}
  1. Returns 寿司, susi (Kunrei-shiki) and sushi (Hepburn-shiki).

The kuromoji_completion token filter accepts the following settings:

mode
The tokenization mode determines how the tokenizer handles compound and unknown words. It can be set to:
index
Simple romanization. Expected to be used when indexing.
query
Input Method aware romanization. Expected to be used when querying.

Defaults to index.