I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
Thanks for your answer, I tried using the scan search type but the
behavior is the same, the curl request against the search_id hangs
forever and the elasticsearch node against which the query was run
becomes non-responsive...
Thanks
On Jan 17, 11:17 am, Berkay Mollamustafaoglu mber...@gmail.com
wrote:
Hi,
You may want to take a look at the scan search typehttp://www.elasticsearch.org/guide/reference/api/search/search-type.html
Regards,
Berkay Mollamustafaoglu
mberkay on yahoo, google and skype
I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
On Tue, 2012-01-17 at 13:43 -0800, Simeon Zaharici wrote:
Hello
Thanks for your answer, I tried using the scan search type but the
behavior is the same, the curl request against the search_id hangs
forever and the elasticsearch node against which the query was run
becomes non-responsive...
You didn't specify what value $MESSAGES has.
The idea is to use search_type=scan and to use a scrolled search, with a
reasonable size (eg 1000).
So you pull the first 1000 (x no of primary shards), then keep pulling
until there are no more records left to pull
clint
Thanks
On Jan 17, 11:17 am, Berkay Mollamustafaoglu mber...@gmail.com
wrote:
Hi,
You may want to take a look at the scan search typehttp://www.elasticsearch.org/guide/reference/api/search/search-type.html
Regards,
Berkay Mollamustafaoglu
mberkay on yahoo, google and skype
I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
I would like to be able to extract all log messages that are created in one day for archiving purposes.
Why not backup one full index? It would be much faster and cheaper in
terms of CPU? (not sure if graylog has an option to move to a new
index per day or sth.)
I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
On Tue, 2012-01-17 at 13:43 -0800, Simeon Zaharici wrote:
Hello
Thanks for your answer, I tried using the scan search type but the
behavior is the same, the curl request against the search_id hangs
forever and the elasticsearch node against which the query was run
becomes non-responsive...
You didn't specify what value $MESSAGES has.
The idea is to use search_type=scan and to use a scrolled search, with a
reasonable size (eg 1000).
So you pull the first 1000 (x no of primary shards), then keep pulling
until there are no more records left to pull
clint
Thanks
On Jan 17, 11:17 am, Berkay Mollamustafaoglu mber...@gmail.com
wrote:
Hi,
You may want to take a look at the scan search typehttp://www.elasticsearch.org/guide/reference/api/search/search-type.html
Regards,
Berkay Mollamustafaoglu
mberkay on yahoo, google and skype
I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
I would like to be able to extract all log messages that are created in one day for archiving purposes.
Why not backup one full index? It would be much faster and cheaper in
terms of CPU? (not sure if graylog has an option to move to a new
index per day or sth.)
I am a new user of elasticsearch. We are using graylog2, a log
management solution that stores the log messages in elasticsearch.
I would like to be able to extract all log messages that are created
in one day for archiving purposes. Our servers and applications
generate around 5 million entries a day. What would be the best way to
extract these entries from elasticsearch ?
When doing a search that would match all these messages the query
hangs forever and the load on the elasticserch node against which the
query is executed goes up.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.