We are running Marvel 1.2.1 on ES 1.3.4/Java 1.7.55 with a separate
"logging cluster" that marvel.agent.exporter.es.hosts points to. The
logging cluster is running the same bits as our production cluster
with marvel.agent.enabled set to false. About a week ago, we stopped
seeing events for the "Cluster Pulse" and "Shard Allocation" dashboards.
Whenever one of the events tracked on these dashboards occurs, the current
master spits out the following log message:
[2014-12-09 02:54:59,220][ERROR][marvel.agent.exporter ] [T02-C01-M03]
remote target didn't respond with 200 OK response code [500 Internal Server
Error]. content: [
?error?JsonParseException[Invalid shared name reference 293; only got 0
names in buffer (invalid content)
at [Source:
org.elasticsearch.transport.netty.ChannelBufferStreamInput@3fc84b46; line:
-1, column: 4]]?status$ ??]
On the logging cluster, the relevant types for these pages (node_event,
shard_event, etc) dried up at the same time the errors started appearing in
the logs. I checked the both the Marvel template and the actual mappings -
no differences between the two days. Has anyone seen a similar issue?
On Tuesday, December 9, 2014 4:11:00 AM UTC+1, Lane Harris wrote:
Hello all
We are running Marvel 1.2.1 on ES 1.3.4/Java 1.7.55 with a separate
"logging cluster" that marvel.agent.exporter.es.hosts points to. The
logging cluster is running the same bits as our production cluster
with marvel.agent.enabled set to false. About a week ago, we stopped
seeing events for the "Cluster Pulse" and "Shard Allocation" dashboards.
Whenever one of the events tracked on these dashboards occurs, the current
master spits out the following log message:
[2014-12-09 02:54:59,220][ERROR][marvel.agent.exporter ] [T02-C01-M03]
remote target didn't respond with 200 OK response code [500 Internal Server
Error]. content: [
?error?JsonParseException[Invalid shared name reference 293; only got 0
names in buffer (invalid content)
at [Source:
org.elasticsearch.transport.netty.ChannelBufferStreamInput@3fc84b46; line:
-1, column: 4]]?status$ ??]
On the logging cluster, the relevant types for these pages (node_event,
shard_event, etc) dried up at the same time the errors started appearing in
the logs. I checked the both the Marvel template and the actual mappings -
no differences between the two days. Has anyone seen a similar issue?
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.