fix #2416 add standard_search_analyzer

This commit is contained in:
Shinsuke Sugaya 2020-02-29 05:52:32 +09:00
parent 91bf1e577c
commit 59e07f6bb6
3 changed files with 30 additions and 2 deletions

View file

@ -18,7 +18,7 @@
<param name="plugin.name.prefix" value="elasticsearch-" />
<param name="plugin.name" value="analysis-extension" />
<param name="plugin.version" value="7.6.1-SNAPSHOT" />
<param name="plugin.zip.version" value="7.6.1-20200227.133255-8" />
<param name="plugin.zip.version" value="7.6.1-20200228.143157-9" />
</antcall>
<!-- analysis-fess -->
<antcall target="install.plugin">

View file

@ -685,6 +685,15 @@
"expand_ngram": true,
"reload_interval": "1m"
},
"unigram_search_tokenizer": {
"type": "ngram_synonym",
"n": "1",
"synonyms_path": "${fess.dictionary.path}synonym.txt",
"dynamic_reload": true,
"expand": false,
"expand_ngram": false,
"reload_interval": "1m"
},
"bigram_tokenizer": {
"type": "ngram_synonym",
"n": "2",
@ -1145,6 +1154,23 @@
"stemmer_en_filter"
]
},
"standard_search_analyzer": {
"type": "custom",
"char_filter": [
"mapping_filter"
],
"tokenizer": "unigram_search_tokenizer",
"filter": [
"disable_graph",
"alphanum_word_filter",
"cjk_bigram",
"lowercase",
"stopword_en_filter",
"english_keywords",
"english_override",
"stemmer_en_filter"
]
},
"empty_analyzer": {
"type": "custom",
"tokenizer": "standard",
@ -1157,7 +1183,7 @@
"char_filter": [
"mapping_ja_filter"
],
"tokenizer": "unigram_tokenizer",
"tokenizer": "unigram_search_tokenizer",
"filter": [
"alphanum_word_filter",
"cjk_bigram",

View file

@ -498,6 +498,7 @@
"content": {
"type": "text",
"analyzer": "standard_analyzer",
"search_analyzer": "standard_search_analyzer",
"term_vector": "with_positions_offsets"
},
"content_minhash": {
@ -576,6 +577,7 @@
"title": {
"type": "text",
"analyzer": "standard_analyzer",
"search_analyzer": "standard_search_analyzer",
"term_vector": "with_positions_offsets"
},
"thumbnail": {