I have two servers running an elasticsearch cluster as our website's search
engine. And use Elastica as our php client.
At beginning, the queries are sent directly to ES but the servers are very
unstable and the tcp connections are about 500-600, so ES can't handle them
quickly and always get timeout response (We set the timeout to 5s). So we
added 5mins cache with memcached and the situation got better. The tcp
connections are controlled around 10 (avg).
I found that if the connections over 100 then it will become very unstable.
Does this because the server can't handle too much request? Or I need to
optimize my queries? (Most queries took about 50 ms)
I have two servers running an elasticsearch cluster as our website's search
engine. And use Elastica as our php client.
At beginning, the queries are sent directly to ES but the servers are very
unstable and the tcp connections are about 500-600, so ES can't handle them
quickly and always get timeout response (We set the timeout to 5s). So we
added 5mins cache with memcached and the situation got better. The tcp
connections are controlled around 10 (avg).
I found that if the connections over 100 then it will become very unstable.
Does this because the server can't handle too much request? Or I need to
optimize my queries? (Most queries took about 50 ms)
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
So I still seem to be having the same issue. I have two machines with
16GB RAM each on them. A 10,000 RPM drive. As the datasize increased
to about 20GB total, 20 million documents, the queries seem to be
taking longer and the connections start to backup until it no longer
seems to be taking HTTP requests.
The logs show nothing. There are no huge CPU usage, or heap usage,
just dead. Any ideas on what I can paste here in terms of logs to
debug the issue?
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
So I still seem to be having the same issue. I have two machines with
16GB RAM each on them. A 10,000 RPM drive. As the datasize increased
to about 20GB total, 20 million documents, the queries seem to be
taking longer and the connections start to backup until it no longer
seems to be taking HTTP requests.
The logs show nothing. There are no huge CPU usage, or heap usage,
just dead. Any ideas on what I can paste here in terms of logs to
debug the issue?
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
So I still seem to be having the same issue. I have two machines with
16GB RAM each on them. A 10,000 RPM drive. As the datasize increased
to about 20GB total, 20 million documents, the queries seem to be
taking longer and the connections start to backup until it no longer
seems to be taking HTTP requests.
The logs show nothing. There are no huge CPU usage, or heap usage,
just dead. Any ideas on what I can paste here in terms of logs to
debug the issue?
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
So I still seem to be having the same issue. I have two machines with
16GB RAM each on them. A 10,000 RPM drive. As the datasize increased
to about 20GB total, 20 million documents, the queries seem to be
taking longer and the connections start to backup until it no longer
seems to be taking HTTP requests.
The logs show nothing. There are no huge CPU usage, or heap usage,
just dead. Any ideas on what I can paste here in terms of logs to
debug the issue?
I notice when my total index size exceeds the RAM size (16GB of ram
per machine) the queries start to take a bit longer. Once the
connections pile up the entire cluster becomes massively unstable and
crashes. I have a theory as the dataset goes up in size what was once
a fast query suddenly is slow (my queries fetch data from a certain
time and sort) and I think that might be killing the cluster.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.