The stop analyzer is the same as the simple analyzer
but adds support for removing stop words. It defaults to using the
_english_ stop words.
Example output
POST _analyze
{
"analyzer": "stop",
"text": "The 2 QUICK Brown-Foxes jumped over the lazy dog's bone."
}
The above sentence would produce the following terms:
[ quick, brown, foxes, jumped, over, lazy, dog, s, bone ]
Configuration
The stop analyzer accepts the following parameters:
|
|
A pre-defined stop words list like |
|
|
The path to a file containing stop words. This path is relative to the
Elasticsearch |
See the Stop Token Filter for more information about stop word configuration.
Example configuration
In this example, we configure the stop analyzer to use a specified list of
words as stop words:
PUT my-index-000001
{
"settings": {
"analysis": {
"analyzer": {
"my_stop_analyzer": {
"type": "stop",
"stopwords": ["the", "over"]
}
}
}
}
}
POST my-index-000001/_analyze
{
"analyzer": "my_stop_analyzer",
"text": "The 2 QUICK Brown-Foxes jumped over the lazy dog's bone."
}
The above example produces the following terms:
[ quick, brown, foxes, jumped, lazy, dog, s, bone ]
Definition
It consists of:
- Tokenizer
- Token filters
If you need to customize the stop analyzer beyond the configuration
parameters then you need to recreate it as a custom analyzer and modify
it, usually by adding token filters. This would recreate the built-in
stop analyzer and you can use it as a starting point for further
customization: