Hi, By removing all my translog files, ES can start without error.
On Wednesday, January 14, 2015 at 2:56:48 PM UTC+1, Thomas Cataldo wrote: > > Hi, > > I encounter a problem with a large index (38GB) that prevents ES 1.4.2 > from starting. > The problem looks pretty similar to the one in > https://github.com/elasticsearch/elasticsearch/issues/8394 > > I tried some of the recommandations from this post (and linked ones) : > > index.load_fixed_bitset_filters_eagerly: false > index.warmer.enabled: false > indices.breaker.total.limit: 30% > > And event with that, my server does not start [1]. > > I uploaded to gist the mapping for the index : > https://gist.github.com/tcataldo/c0b6b3dfec9823bf6523 > > I tried several OS memory, ES heap combinations, the biggest being > 48GiB for the operating system and 32GiB for ES heap and it still > fails with that. > > Any idea or link to an open issue I could follow ? > > Regards, > Thomas. > > > > 1. debug output: > > [2015-01-14 12:01:55,740][DEBUG][indices.cluster ] [Saint Elmo] > [mailspool][0] creating shard > [2015-01-14 12:01:55,741][DEBUG][index.service ] [Saint Elmo] > [mailspool] creating shard_id [0] > [2015-01-14 12:01:56,041][DEBUG][index.deletionpolicy ] [Saint Elmo] > [mailspool][0] Using [keep_only_last] deletion policy > [2015-01-14 12:01:56,041][DEBUG][index.merge.policy ] [Saint Elmo] > [mailspool][0] using [tiered] merge mergePolicy with > expunge_deletes_allowed[10.0], floor_segment[2mb], max_merge_at_on\ > ce[10], max_merge_at_once_explicit[30], max_merged_segment[5gb], > segments_per_tier[10.0], reclaim_deletes_weight[2.0] > [2015-01-14 12:01:56,041][DEBUG][index.merge.scheduler ] [Saint Elmo] > [mailspool][0] using [concurrent] merge scheduler with max_thread_count[2], > max_merge_count[4] > [2015-01-14 12:01:56,042][DEBUG][index.shard.service ] [Saint Elmo] > [mailspool][0] state: [CREATED] > [2015-01-14 12:01:56,043][DEBUG][index.translog ] [Saint Elmo] > [mailspool][0] interval [5s], flush_threshold_ops [2147483647], > flush_threshold_size [200mb], flush_threshold_period [3\ > 0m] > [2015-01-14 12:01:56,044][DEBUG][index.shard.service ] [Saint Elmo] > [mailspool][0] state: [CREATED]->[RECOVERING], reason [from gateway] > [2015-01-14 12:01:56,044][DEBUG][index.gateway ] [Saint Elmo] > [mailspool][0] starting recovery from local ... > [2015-01-14 12:01:56,048][DEBUG][river.cluster ] [Saint Elmo] > processing [reroute_rivers_node_changed]: execute > [2015-01-14 12:01:56,048][DEBUG][river.cluster ] [Saint Elmo] > processing [reroute_rivers_node_changed]: no change in cluster_state > [2015-01-14 12:01:56,048][DEBUG][cluster.service ] [Saint Elmo] > processing [shard-failed ([mailspool][3], node[gOgAuHo4SXyfyuPpws0Usw], > [P], s[INITIALIZING]), reason [engine failure, \ > message [refresh failed][OutOfMemoryError[Java heap space]]]]: done > applying updated cluster_state (version: 4) > [2015-01-14 12:01:56,062][DEBUG][index.engine.internal ] [Saint Elmo] > [mailspool][0] starting engine > [2015-01-14 12:02:19,701][WARN ][index.engine.internal ] [Saint Elmo] > [mailspool][0] failed engine [refresh failed] > java.lang.OutOfMemoryError: Java heap space > at org.apache.lucene.util.FixedBitSet.<init>(FixedBitSet.java:187) > at > org.apache.lucene.search.MultiTermQueryWrapperFilter.getDocIdSet(MultiTermQueryWrapperFilter.java:104) > at > org.elasticsearch.index.cache.filter.weighted.WeightedFilterCache$FilterCacheFilterWrapper.getDocIdSet(WeightedFilterCache.java:177) > at > org.elasticsearch.common.lucene.search.OrFilter.getDocIdSet(OrFilter.java:55) > at > org.elasticsearch.common.lucene.search.ApplyAcceptedDocsFilter.getDocIdSet(ApplyAcceptedDocsFilter.java:46) > at > org.apache.lucene.search.FilteredQuery$1.scorer(FilteredQuery.java:130) > at > org.apache.lucene.search.FilteredQuery$RandomAccessFilterStrategy.filteredScorer(FilteredQuery.java:542) > at > org.apache.lucene.search.FilteredQuery$1.scorer(FilteredQuery.java:136) > at > org.apache.lucene.search.QueryWrapperFilter$1.iterator(QueryWrapperFilter.java:59) > at > org.apache.lucene.index.BufferedUpdatesStream.applyQueryDeletes(BufferedUpdatesStream.java:554) > at > org.apache.lucene.index.BufferedUpdatesStream.applyDeletesAndUpdates(BufferedUpdatesStream.java:287) > at > org.apache.lucene.index.IndexWriter.applyAllDeletesAndUpdates(IndexWriter.java:3271) > at > org.apache.lucene.index.IndexWriter.maybeApplyDeletes(IndexWriter.java:3262) > at > org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:421) > at > org.apache.lucene.index.StandardDirectoryReader.doOpenFromWriter(StandardDirectoryReader.java:292) > at > org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:267) > at > org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:257) > at > org.apache.lucene.index.DirectoryReader.openIfChanged(DirectoryReader.java:171) > at > org.apache.lucene.search.SearcherManager.refreshIfNeeded(SearcherManager.java:118) > at > org.apache.lucene.search.SearcherManager.refreshIfNeeded(SearcherManager.java:58) > at > org.apache.lucene.search.ReferenceManager.doMaybeRefresh(ReferenceManager.java:176) > at > org.apache.lucene.search.ReferenceManager.maybeRefresh(ReferenceManager.java:225) > at > org.elasticsearch.index.engine.internal.InternalEngine.refresh(InternalEngine.java:796) > at > org.elasticsearch.index.engine.internal.InternalEngine.delete(InternalEngine.java:692) > at > org.elasticsearch.index.shard.service.InternalIndexShard.performRecoveryOperation(InternalIndexShard.java:798) > at > org.elasticsearch.index.gateway.local.LocalIndexShardGateway.recover(LocalIndexShardGateway.java:268) > at > org.elasticsearch.index.gateway.IndexShardGatewayService$1.run(IndexShardGatewayService.java:132) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > [2015-01-14 12:02:19,704][DEBUG][index.service ] [Saint Elmo] > [mailspool] [0] closing... (reason: [engine failure, message [refresh > failed][OutOfMemoryError[Java heap space]]]) > > -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/91792194-7511-4d16-992b-67b23f6c8bb0%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.