This bug occurs because you are upgrading to an old version of elasticsearch (1.3.4). Try the latest version where the bug is fixed: https://issues.apache.org/jira/browse/LUCENE-5975
On Fri, Dec 19, 2014 at 5:40 AM, Georgeta Boanea <gio...@gmail.com> wrote: > Hi All, > > After upgrading from ES 0.20.6 to 1.3.4 the following messages occurred: > > [2014-12-19 10:02:06.714 GMT] WARN |||||| > elasticsearch[es-node-name][generic][T#14] > org.elasticsearch.cluster.action.shard [es-node-name] [index-name][3] > sending failed shard for [index-name][3], node[qOTLmb3IQC2COXZh1n9O2w], [P], > s[INITIALIZING], indexUUID [_na_], reason [Failed to start shard, message > [IndexShardGatewayRecoveryException[[index-name][3] failed to fetch index > version after copying it over]; nested: > CorruptIndexException[[index-name][3] Corrupted index > [corrupted_Ackui00SSBi8YXACZGNDkg] caused by: CorruptIndexException[did not > read all bytes from file: read 112 vs size 113 (resource: > BufferedChecksumIndexInput(NIOFSIndexInput(path="path/3/index/_uzm_2.del")))]]; > ]] > > [2014-12-19 10:02:08.390 GMT] WARN |||||| > elasticsearch[es-node-name][generic][T#20] org.elasticsearch.indices.cluster > [es-node-name] [index-name][3] failed to start shard > org.elasticsearch.index.gateway.IndexShardGatewayRecoveryException: > [index-name][3] failed to fetch index version after copying it over > at > org.elasticsearch.index.gateway.local.LocalIndexShardGateway.recover(LocalIndexShardGateway.java:152) > at > org.elasticsearch.index.gateway.IndexShardGatewayService$1.run(IndexShardGatewayService.java:132) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > Caused by: org.apache.lucene.index.CorruptIndexException: [index-name][3] > Corrupted index [corrupted_Ackui00SSBi8YXACZGNDkg] caused by: > CorruptIndexException[did not read all bytes from file: read 112 vs size 113 > (resource: > BufferedChecksumIndexInput(NIOFSIndexInput(path="path/3/index/_uzm_2.del")))] > at org.elasticsearch.index.store.Store.failIfCorrupted(Store.java:353) > at org.elasticsearch.index.store.Store.failIfCorrupted(Store.java:338) > at > org.elasticsearch.index.gateway.local.LocalIndexShardGateway.recover(LocalIndexShardGateway.java:119) > ... 4 more > > Shard [3] of the index remains unallocated and the cluster remains in a RED > state. > > curl -XGET 'http://localhost:48012/_cluster/health?pretty=true' > { > "cluster_name" : "cluster-name", > "status" : "red", > "timed_out" : false, > "number_of_nodes" : 5, > "number_of_data_nodes" : 5, > "active_primary_shards" : 10, > "active_shards" : 20, > "relocating_shards" : 0, > "initializing_shards" : 1, > "unassigned_shards" : 1 > } > > If I do an optimize (curl -XPOST > http://localhost:48012/index-name/_optimize?max_num_segments=1) for the > index before the update, everything is fine. Optimize works just before the > update, if is done after the update the problem remains the same. > > Any idea why this problem occurs? > Is there another way to avoid this problem? I want to avoid optimize in case > of large volume of data. > > Thank you, > Georgeta > > -- > You received this message because you are subscribed to the Google Groups > "elasticsearch" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to elasticsearch+unsubscr...@googlegroups.com. > To view this discussion on the web visit > https://groups.google.com/d/msgid/elasticsearch/74d0af86-c661-4e58-ba2c-d38adde1291c%40googlegroups.com. > For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/CAMUKNZW3Kc-8smWQjn1VRrk2yhgdiA33EctWUiXEOkxg46BjiQ%40mail.gmail.com. For more options, visit https://groups.google.com/d/optout.