We running version 0.19.9 with 6 servers running using 6 shards. A few
days ago, shard number 2 seems to have gone goofy (looks like
a corruption in the index) causing the following exception to appear
constantly in the server logs:
org.elasticsearch.transport.RemoteTransportException:
[cardano][inet[/xx.xxx.xx.xxx:9300]][search/phase/query] Caused by: org.elasticsearch.search.query.QueryPhaseExecutionException:
[theindex][2]: query[filtered(+activityObject.content:"Some query term"
+sourceInfo.publisher:Some Name
-sourceInfo.dataSource:directPooling)->cache(_type:socialmedia)],from[0],size[1],sort[<custom:"sortDate":
org.elasticsearch.index.field.data.longs.LongFieldDataType$1@24f99e97>!]:
Query Failed [Failed to execute main query]
at org.elasticsearch.search.query.QueryPhase.execute(QueryPhase.java:182)*
at
org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:234)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:497)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:486)
at
org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:268)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)*
*Caused by: java.io.EOFException: read past EOF:
NIOFSIndexInput(path="/var/data/elasticsearch/nodes/0/indices/theindex/2/index/_161lvl.tis")
at
org.apache.lucene.store.BufferedIndexInput.refill(BufferedIndexInput.java:264)
at
org.apache.lucene.store.BufferedIndexInput.readByte(BufferedIndexInput.java:40)
at org.apache.lucene.store.DataInput.readVInt(DataInput.java:107)*
at
org.apache.lucene.store.BufferedIndexInput.readVInt(BufferedIndexInput.java:217)
What I've tried:
Changed replication to 0
Closed/Opened the index (to force rebalancing)
Restarted the node containing shard 2
with index.shard.check_on_startup: true
This seems to point to something going bad with the lucene index of this
shard, you the check_on_startup didn't seem to solve the problem. Anyone
know how to get around this.
We running version 0.19.9 with 6 servers running using 6 shards. A few
days ago, shard number 2 seems to have gone goofy (looks like
a corruption in the index) causing the following exception to appear
constantly in the server logs:
org.elasticsearch.transport.RemoteTransportException:
[cardano][inet[/xx.xxx.xx.xxx:9300]][search/phase/query] Caused by: org.elasticsearch.search.query.QueryPhaseExecutionException:
[theindex][2]: query[filtered(+activityObject.content:"Some query term"
+sourceInfo.publisher:Some Name
-sourceInfo.dataSource:directPooling)->cache(_type:socialmedia)],from[0],size[1],sort[<custom:"sortDate":
org.elasticsearch.index.field.data.longs.LongFieldDataType$1@24f99e97>!]:
Query Failed [Failed to execute main query]
at
org.elasticsearch.search.query.QueryPhase.execute(QueryPhase.java:182)*
at
org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:234)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:497)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:486)
at
org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:268)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)*
*Caused by: java.io.EOFException: read past EOF:
NIOFSIndexInput(path="/var/data/elasticsearch/nodes/0/indices/theindex/2/index/_161lvl.tis")
at
org.apache.lucene.store.BufferedIndexInput.refill(BufferedIndexInput.java:264)
at
org.apache.lucene.store.BufferedIndexInput.readByte(BufferedIndexInput.java:40)
at org.apache.lucene.store.DataInput.readVInt(DataInput.java:107)*
at
org.apache.lucene.store.BufferedIndexInput.readVInt(BufferedIndexInput.java:217)
What I've tried:
Changed replication to 0
Closed/Opened the index (to force rebalancing)
Restarted the node containing shard 2
with index.shard.check_on_startup: true
This seems to point to something going bad with the lucene index of this
shard, you the check_on_startup didn't seem to solve the problem. Anyone
know how to get around this.
On Fri, Jan 4, 2013 at 11:36 AM, Dimitry <dlvo...@gmail.com <javascript:>>wrote:
We running version 0.19.9 with 6 servers running using 6 shards. A few
days ago, shard number 2 seems to have gone goofy (looks like
a corruption in the index) causing the following exception to appear
constantly in the server logs:
org.elasticsearch.transport.RemoteTransportException:
[cardano][inet[/xx.xxx.xx.xxx:9300]][search/phase/query] Caused by: org.elasticsearch.search.query.QueryPhaseExecutionException:
[theindex][2]: query[filtered(+activityObject.content:"Some query term"
+sourceInfo.publisher:Some Name
-sourceInfo.dataSource:directPooling)->cache(_type:socialmedia)],from[0],size[1],sort[<custom:"sortDate":
org.elasticsearch.index.field.data.longs.LongFieldDataType$1@24f99e97>!]:
Query Failed [Failed to execute main query]
at
org.elasticsearch.search.query.QueryPhase.execute(QueryPhase.java:182)*
at
org.elasticsearch.search.SearchService.executeQueryPhase(SearchService.java:234)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:497)
at
org.elasticsearch.search.action.SearchServiceTransportAction$SearchQueryTransportHandler.messageReceived(SearchServiceTransportAction.java:486)
at
org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:268)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)*
*Caused by: java.io.EOFException: read past EOF:
NIOFSIndexInput(path="/var/data/elasticsearch/nodes/0/indices/theindex/2/index/_161lvl.tis")
at
org.apache.lucene.store.BufferedIndexInput.refill(BufferedIndexInput.java:264)
at
org.apache.lucene.store.BufferedIndexInput.readByte(BufferedIndexInput.java:40)
at org.apache.lucene.store.DataInput.readVInt(DataInput.java:107)*
at
org.apache.lucene.store.BufferedIndexInput.readVInt(BufferedIndexInput.java:217)
What I've tried:
Changed replication to 0
Closed/Opened the index (to force rebalancing)
Restarted the node containing shard 2
with index.shard.check_on_startup: true
This seems to point to something going bad with the lucene index of this
shard, you the check_on_startup didn't seem to solve the problem. Anyone
know how to get around this.
Apache, Apache Lucene, Apache Hadoop, Hadoop, HDFS and the yellow elephant
logo are trademarks of the
Apache Software Foundation
in the United States and/or other countries.