Hi,
we use elasticsearch for storing some metadata of CI, we have automation's to collect metadata and push to elasticsearch.
Our elasticsearch cluster has below configuration,
2 - master nodes
3 - data nodes
all nodes are individual VM's with different IP's connected into 1 cluster
We configure one of the active master node IP in metadata collectors to push data to ES,
But our question is what if that node goes down for some reason ? We don't want to reconfigure our collectors with new master node IP.
Is there any best way to handle such scenarios ? I guess its quite common scenario when working with clusters
as of now no, collectors doesn't support load balancing or multiple hosts configuration.
We are looking if this can be done in some other way than collector code change (this we want to handle as a last option)
Collects data from our various CI sources,
One is maven plugin: to collect data of Maven artifacts generated in maven build.
Jenkins plugin: to collect non-maven CI data.
No, not by parsing by logs.
they have different logic, example maven plugin uses maven-core API's and loop through all the projects built and generate report.
ok.
for redundancy it follows different way, it writes data into a file and upload to binary repo. When ES is UP later, We have automation which reads these files and push data to ES.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.