if I group by the field, I will need to create transform for each field (I think), which means every transform will look for new data every time interval (lets say 1h), suppose I have 5 fields, then its 5 transforms => 5 queries every 1h. Am I right ?
here is a data example:
the raw data:
let's call it client_data index
{
"utm_campaign" : "none",
"utm_medium" : "none",
"utm_source" : "none",
"fullUrl" : "....",
"clientId" : "100",
"user_id" : "JD1GODDi85x1kH1uG32j7e71iEz630",
"country" : "Israel",
"wz_session" : "JD1GODi85x1kH1u32j7e71iEz630",
"date_time" : "2021-02-09T11:52:02",
"protocol" : "https",
"domain" : "elastic.co",
"page" : "....",
"raw_page" : "....",
"operating_system" : "Android 11",
"browser" : "Chrome Mobile",
"browser_version" : "88.0.4324.152",
"is_bot" : false,
"device" : "Phone",
}
I'm looking for a document per unique combination, because if I keep all the unique values in 1 document I could end up with a document with a field containing thousands of unique values.
so I'm looking for something like this:
let's call it client_lookups index
{
"clientId" : 100,
"domain" : "elastic.co"
"page" : "/this_this_page_path", // or device, country, browser, browser version, utm...
"count" : 3, // not mandatory, but nice to have it
"lastHitDate" : "2021-07-07T07:30:00", // not mandatory, but nice to have it
}
I thought about it again now, and I think there is no way to do that with 1 transform (if I pivot to group by: clientId, doman, page, value). the question is if there is a job that looks for a new data in an index and check if any of the fields [page, country,...] has a new unique value.