I know that you can configure an analyzer per field via _mapping, but if I
want to configure an analyzer or tokenizer for an entire schema
(unfortunately, I do not know the schema upfront), is there a way? Any
examples will be highly appreciated.
The real problem I am trying to solve is avoid tokenization for some of the
fileds that I need to facet on. I need the facets to count the exact phrase
and not tokenized pieces. However, there are tons of fields that fall into
the above and I do not know want to configure each such attribute
individually.
You can change the default analyzer at the index level [1]. "Schemas" tend
to be organized per type, but since each similarly named field is shared in
the index, each field should have the same analysis.
Of course, the default analyzer will be used on every field. You can change
the analyzer per field if you match on certain attributes such as name or
type [2].
I know that you can configure an analyzer per field via _mapping, but if I
want to configure an analyzer or tokenizer for an entire schema
(unfortunately, I do not know the schema upfront), is there a way? Any
examples will be highly appreciated.
The real problem I am trying to solve is avoid tokenization for some of
the fileds that I need to facet on. I need the facets to count the exact
phrase and not tokenized pieces. However, there are tons of fields that
fall into the above and I do not know want to configure each such attribute
individually.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.