logstash/docs/reference/field-extraction.md
Karen Metts 91927d7450
Doc: Migrate docs from AsciiDoc to Markdown in 9.0 branch (#17289)
* Doc: Delete asciidoc files for 9.0 branch
* Add MD files for 9.0 branch
2025-03-10 18:02:14 -04:00

3.2 KiB
Raw Blame History

mapped_pages
https://www.elastic.co/guide/en/logstash/current/field-extraction.html

Extracting Fields and Wrangling Data [field-extraction]

The plugins described in this section are useful for extracting fields and parsing unstructured data into fields.

dissect filter
Extracts unstructured event data into fields by using delimiters. The dissect filter does not use regular expressions and is very fast. However, if the structure of the data varies from line to line, the grok filter is more suitable.

For example, lets say you have a log that contains the following message:

Apr 26 12:20:02 localhost systemd[1]: Starting system activity accounting tool...

The following config dissects the message:

filter {
  dissect {
    mapping => { "message" => "%{ts} %{+ts} %{+ts} %{src} %{prog}[%{pid}]: %{msg}" }
  }
}

After the dissect filter is applied, the event will be dissected into the following fields:

{
  "msg"        => "Starting system activity accounting tool...",
  "@timestamp" => 2017-04-26T19:33:39.257Z,
  "src"        => "localhost",
  "@version"   => "1",
  "host"       => "localhost.localdomain",
  "pid"        => "1",
  "message"    => "Apr 26 12:20:02 localhost systemd[1]: Starting system activity accounting tool...",
  "type"       => "stdin",
  "prog"       => "systemd",
  "ts"         => "Apr 26 12:20:02"
}
kv filter
Parses key-value pairs.

For example, lets say you have a log message that contains the following key-value pairs:

ip=1.2.3.4 error=REFUSED

The following config parses the key-value pairs into fields:

filter {
  kv { }
}

After the filter is applied, the event in the example will have these fields:

  • ip: 1.2.3.4
  • error: REFUSED
grok filter
Parses unstructured event data into fields. This tool is perfect for syslog logs, Apache and other webserver logs, MySQL logs, and in general, any log format that is generally written for humans and not computer consumption. Grok works by combining text patterns into something that matches your logs.

For example, lets say you have an HTTP request log that contains the following message:

55.3.244.1 GET /index.html 15824 0.043

The following config parses the message into fields:

filter {
  grok {
    match => { "message" => "%{IP:client} %{WORD:method} %{URIPATHPARAM:request} %{NUMBER:bytes} %{NUMBER:duration}" }
  }
}

After the filter is applied, the event in the example will have these fields:

  • client: 55.3.244.1
  • method: GET
  • request: /index.html
  • bytes: 15824
  • duration: 0.043

::::{tip} If you need help building grok patterns, try the Grok Debugger. The Grok Debugger is an {{xpack}} feature under the Basic License and is therefore free to use. ::::