That I know, I want to automate this process because I have 10+ space and each has five-six dashboards and three different clusters
means 150 dashboards.
Following command works from bash script and I jsut have to get all the ids for all dashboard. because in my cluster user has access to create what they want and any given day user might create new dashboard. I would like to get that backed up automatically.
You will need a tool like jq to manipulate JSON responses and iterate through them.
The best way to backup dashboards is to use the Snapshot/Restore capabilities of Elasticsearch, just snapshotting all the indices named prefixed .kibana*.
Something like this will search all the dashboards and generate one file per dashboard.
If you have multiple spaces, you'll need to do another external loop to search for spaces and insert the space id in the URL.
Reason I want to save seperate dashboard is because if someone mess up his/her own dashboard I can restore that one. by just backing up whole .kibana* indexes I can't do that.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.