elasticsearch/docs/reference/elasticsearch-plugins/analysis-kuromoji-completion.md
Colleen McGinnis b7e3a1e14b
[docs] Migrate docs from AsciiDoc to Markdown (#123507)
* delete asciidoc files

* add migrated files

* fix errors

* Disable docs tests

* Clarify release notes page titles

* Revert "Clarify release notes page titles"

This reverts commit 8be688648d.

* Comment out edternal URI images

* Clean up query languages landing pages, link to conceptual docs

* Add .md to url

* Fixes inference processor nesting.

---------

Co-authored-by: Liam Thompson <32779855+leemthompo@users.noreply.github.com>
Co-authored-by: Liam Thompson <leemthompo@gmail.com>
Co-authored-by: Martijn Laarman <Mpdreamz@gmail.com>
Co-authored-by: István Zoltán Szabó <szabosteve@gmail.com>
2025-02-27 17:56:14 +01:00

34 lines
861 B
Markdown

---
mapped_pages:
- https://www.elastic.co/guide/en/elasticsearch/plugins/current/analysis-kuromoji-completion.html
---
# kuromoji_completion token filter [analysis-kuromoji-completion]
The `kuromoji_completion` token filter adds Japanese romanized tokens to the term attributes along with the original tokens (surface forms).
```console
GET _analyze
{
"analyzer": "kuromoji_completion",
"text": "寿司" <1>
}
```
1. Returns `寿司`, `susi` (Kunrei-shiki) and `sushi` (Hepburn-shiki).
The `kuromoji_completion` token filter accepts the following settings:
`mode`
: The tokenization mode determines how the tokenizer handles compound and unknown words. It can be set to:
`index`
: Simple romanization. Expected to be used when indexing.
`query`
: Input Method aware romanization. Expected to be used when querying.
Defaults to `index`.