Combine TokenFilter?

I want a TokenFilter that combines the tokens produced by mutliple other TokenFilters, so that I can create a single field that is tokenized multiple different ways. e.g. a filter section like:

"filter":{
    "prefix_ngram": {
        "min_gram": 1,
        "max_gram": 15,
        "type": "edgeNGram"
    },
    "word_splitter": {
        "type": "pattern_capture",
        "patterns": ["([a-z]+)([0-9.]+)","\\(([^)]+)\\)"],
        "preserve_original": 1
    },
    "all_tokens": {
        "type": "combine",
        "filters": ["prefix_ngram", "word_splitter"],
        "unique": true  
    }
}

I looked but couldn't find anything like this. Is it easy to create?

As a workaround, I've indexed with a document structure that has the same field repeated with multiple different analyzers, and then used a multi_match query, but it would be more elegant / efficient to store all the different tokenizations in a single field.