I do not believe this is possible, at least not through the API. If
you know the average size of each tweet, you could use the Count API and
multiply the count by average doc size.
-Zach
On Sunday, February 10, 2013 9:20:24 AM UTC-5, Elastic Noob wrote:
Hi good day,
I was wondering how can i calculate the amount of disk space used by
Elastic Search?
Here's an example:
I collected 1 million tweets, using the index "tweets" and I indexed the
"text" and "users" key.
So my 2 questions are:
In this situation, how do i find out the amount of disk space used
the the "tweets" index?
Is there any way I can find out the amount of diskspace used by
tweets that contain the hashtag "#YOLO" for example?
I do not believe this is possible, at least not through the API.
If you know the average size of each tweet, you could use the Count API
and multiply the count by average doc size.
-Zach
On Sunday, February 10, 2013 9:20:24 AM UTC-5, Elastic Noob wrote:
Hi good day,
I was wondering how can i calculate the amount of disk space used by
Elastic Search?
Here's an example:
I collected 1 million tweets, using the index "tweets" and I indexed the
"text" and "users" key.
So my 2 questions are:
In this situation, how do i find out the amount of disk space used
the the "tweets" index?
Is there any way I can find out the amount of diskspace used by
tweets that contain the hashtag "#YOLO" for example?
I do not believe this is possible, at least not through the API.
If you know the average size of each tweet, you could use the Count API
and multiply the count by average doc size.
-Zach
On Sunday, February 10, 2013 9:20:24 AM UTC-5, Elastic Noob wrote:
Hi good day,
I was wondering how can i calculate the amount of disk space used by
Elastic Search?
Here's an example:
I collected 1 million tweets, using the index "tweets" and I indexed the
"text" and "users" key.
So my 2 questions are:
In this situation, how do i find out the amount of disk space used
the the "tweets" index?
Is there any way I can find out the amount of diskspace used by
tweets that contain the hashtag "#YOLO" for example?
i was wondering if the size of the index reflect the size of the
total amount of disk space used by the dataset? ( including the raw
_source data ) ?
You want the total -> store -> size(_in_bytes). It should reflect
the total usage of your index's shards across all disks (primaries +
replicas).
Here's a simple way to get this number without looking through the
json.
% curl -s download.elasticsearch.org/es2unix/es >~/bin/es; chmod +x ~/bin/es
% es indices -v wik
status name pri rep size bytes docs
green wiki 5 1 5.3gb 5731796207 753816
On Friday, February 15, 2013 10:56:30 PM UTC+8, Drew Raines wrote:
Elastic Noob wrote:
i was wondering if the size of the index reflect the size of the
total amount of disk space used by the dataset? ( including the raw
_source data ) ?
You want the total -> store -> size(_in_bytes). It should reflect
the total usage of your index's shards across all disks (primaries +
replicas).
Here's a simple way to get this number without looking through the
json.
% curl -s download.elasticsearch.org/es2unix/es >~/bin/es; chmod +x
~/bin/es
% es indices -v wik
status name pri rep size bytes docs
green wiki 5 1 5.3gb 5731796207 753816
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.