Merge tokens (terms) after the tokenisation

Hi there,

I am trying to find a solution to combine all tokens (terms) after tokenisation.

for example - This analyser(my-analyser) produce n tokens after applying "custom_stop" filter. Is there any way to combine all tokens and generate one single token?

I have seen 'fingerprint' filter which combine all tokens but it does sorting as well, which I don't want. Please suggest solution for this.

 `"analysis": {
      "analyzer": {
        "my-analyser": {
          "tokenizer": "standard",
          "filter": [ "custom_stop"]
      "filter": {
        "custom_stop": {
          "type": "stop",
          "ignore_case": true,
          "stopwords": [ "elastic", "aws", "java" ]


Welcome to our community! :smiley:

So if you end up with tokens like ["token0", "token1", "token2"] you want "token0token1token2"?

This topic was automatically closed 28 days after the last reply. New replies are no longer allowed.