My company is about to move to multi-datacenter implementation of Elasticsearch. Each datacenter will have a local cluster that stores the data. We will have tribe nodes (that kibana connects to) in 1 or 2 of the datacenters to combine data from multiple datacenters.
My network team is asking what kind of bandwidth utilization will querying across datacenter take up? Each of the datacenters will have multiple TBs of data. The network team is worried about someone running * query over a month of indices and ES data nodes trying to send that much data over the pipe.
I know there is a limit for recovery in "indices.recovery.max_bytes_per_sec". Does ES cap how much bytes per sec is used for querys?
Thanks for reading