sundar1
(sundar)
September 29, 2016, 9:34am
1
How to get total/average data nodes size through API's?
The below api to get the data nodes allocation each data nodes but i need average or total size of data nodes. Please help me how to get this.
http://localhost:9200/_cat/allocation
shards disk.indices disk.used disk.avail disk.total disk.percent host ip node
29 397.4gb 556.6gb 1.2tb 1.8tb 29 10.236.140.43 10.236.140.43 vc2coma1081153n-node-02
30 469.5gb 1014.2gb 863gb 1.8tb 54 10.240.138.112 10.240.138.112 vc2coma2081152n-node-03
29 445.2gb 545.7gb 1.3tb 1.8tb 29 10.240.148.27 10.240.148.27 vc2coma2085406n-node-05
29 393.9gb 741.4gb 197gb 938.5gb 79 10.240.146.21 10.240.146.21 vc2coma2080685n-node-01
29 395gb 685.2gb 253.3gb 938.5gb 73 10.240.146.45 10.240.146.45 vc2coma2080687n-node-04
try the API _stats
http://localhost:9200/_stats?all=true
look for store size in bytes near the top. (and divide by 1024000000 for Gig. )
Here is my return
{
"_shards": {
"total": 108,
"successful": 108,
"failed": 0
},
"_all": {
"primaries": {
"docs": {
"count": 560336831,
"deleted": 17200
},
"store": {
"size_in_bytes": 257292738493,
"throttle_time_in_millis": 0
},
sundar1
(sundar)
September 30, 2016, 12:03pm
3
Thanks for your reply looks like my store bytes is not matching. Below are my details.
Allocation details :-
shards disk.indices disk.used disk.avail disk.total disk.percent host ip node
31 485.2gb 775.5gb 163gb 938.5gb 82 10.240.146.45 10.240.146.45 vc2coma2080687n-node-04
32 491.7gb 839.5gb 99gb 938.5gb 89 10.240.146.21 10.240.146.21 vc2coma2080685n-node-01
32 569.3gb 670.2gb 1.1tb 1.8tb 35 10.240.148.27 10.240.148.27 vc2coma2085406n-node-05
31 499.1gb 658.3gb 1.1tb 1.8tb 35 10.236.140.43 10.236.140.43 vc2coma1081153n-node-02
32 591.6gb 1.1tb 740.9gb 1.8tb 60 10.240.138.112 10.240.138.112 vc2coma2081152n-node-03
My store info :-
"_shards": {
"total": 158,
"successful": 158,
"failed": 0
},
"_all": {
"primaries": {
"docs": {
"count": 1054938644,
"deleted": 0
},
"store": {
"size_in_bytes": 1410686115558,
"throttle_time_in_millis": 0
}
I calculated disk used column in allocation info. It's coming more than 4 tb but in store showing 1.4 tb. which one is the correct store data.
have you loaded the HQ plugin?
HQ will quickly tell you the size to compare with what the API returns.
For instance one of my clusters shows in HQ
The API returns
1277160888 / 1024000000=1.24G
sundar1
(sundar)
October 4, 2016, 11:41am
5
Thanks Mark for HQ plugin tool. I have installed KOPF plugin in elasticsearch.
Just now i have installed HQ as well. the below are document and size of data.
Store value is not matching with above figures.
Not sure but I think HQ size is only the primaries and not the replicas.
I know in Kopf that the replicas are taken into account. (as seen in your screen cap above. )
Are any indices closed? Closed indices are stored on disk and take up more space but may be included in KPOF stats. I'm not sure?
Anyone else know the difference?