Built-in analyzers
Elasticsearch comes with several analyzers in its standard installation. In the following table, some analyzers are described:
Analyzer |
Description |
---|---|
Standard Analyzer |
This uses Standard Tokenizer to divide text. Other components are Standard Token Filter, Lower Case Token Filter, and Stop Token Filter. It normalizes tokens, lowercases tokens, and also removes unwanted tokens. By default, Elasticsearch applies the standard analyzer. |
Simple Analyzer |
This uses Letter Tokenizer to divide text. Another component is Lower Case Tokenizer. It lowercases tokens. |
Whitespace Analyzer |
This uses Whitespace Tokenizer to divide text at spaces. |
Stop Analyzer |
This uses Letter Tokenizer to divide text. Other components are Lower Case Tokenizer and Stop Token Filter. It removes stop words from token streams. |
Pattern Analyzer |
This uses a regular expression to divide text. It accepts lowercase and stop words setting. |
Language Analyzer |
A set of analyzers analyze the text for a... |