Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-12 Thread joergpra...@gmail.com
I see also exploding numbers in a quick local test on 1.1.0, thousands of
segments are created, but I'm not sure if this behavior is expected or not.
But that does not look like it should. It slows down bulk indexing
significantly.

Issuing an optimize just stalls...

curl '0:9200/test/_optimize?max_num_segments=1'

Does anyone see something similar?

Stack trace excerpt:

elasticsearch[Doctor Strange][optimize][T#1] #81 daemon prio=5 os_prio=31
tid=0x7f85cdd9f800 nid=0xc507 in Object.wait() [0x00012c14]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
at org.apache.lucene.index.IndexWriter.doWait(IndexWriter.java:4339)
- locked 0x0006cee0fad0 (a org.apache.lucene.index.IndexWriter)
at org.apache.lucene.index.IndexWriter.waitForMerges(IndexWriter.java:2262)
- locked 0x0006cee0fad0 (a org.apache.lucene.index.IndexWriter)
at
org.elasticsearch.index.engine.internal.InternalEngine.optimize(InternalEngine.java:998)
at
org.elasticsearch.index.shard.service.InternalIndexShard.optimize(InternalIndexShard.java:599)
at
org.elasticsearch.action.admin.indices.optimize.TransportOptimizeAction.shardOperation(TransportOptimizeAction.java:115)
at
org.elasticsearch.action.admin.indices.optimize.TransportOptimizeAction.shardOperation(TransportOptimizeAction.java:49)
at
org.elasticsearch.action.support.broadcast.TransportBroadcastOperationAction$AsyncBroadcastAction$2.run(TransportBroadcastOperationAction.java:226)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:744)

(node name Doctor Strange was not my selection btw)

Jörg



On Sat, Apr 12, 2014 at 8:09 PM, Thibaut Britz t.br...@trendiction.comwrote:

 Hi,

 We also have this issue in 1.0.2. Per index we have over 1000 segments.
 Optimize doesn't reduce them.

 Would love to see a fix for this.

 Thanks,
 Thibaut



 On Fri, Apr 11, 2014 at 2:22 PM, Nikolas Everett nik9...@gmail.comwrote:

 For reference I'm also on 1.1.0 but I'm not seeing more segments then I
 expect.  I see an average of ~28 per shard on an index I write to
 constantly.  I don't write all that quickly,  50 updates a second.

 Nik


 On Fri, Apr 11, 2014 at 5:46 AM, Adrien Grand 
 adrien.gr...@elasticsearch.com wrote:

 I managed to reproduce the issue locally, I'm looking into it.


 On Fri, Apr 11, 2014 at 9:45 AM, Robin Wallin walro...@gmail.comwrote:

 Hello,

 We are experiencing a related problem with 1.1.0. Segments do not seem
 to merge as they should during indexing. The optimize API does practically
 nothing in terms of lowering the segments count either. The problem
 persists through a cluster restart. The vast amount of segments seem to be
 greatly impact the performance of the cluster, in a very negative way.

 We currently have 414 million documents across 3 nodes, each shard has
 in average 1200 segments(!).

 With 1.0.1 we had even more documents, ~650 million, without any
 segment problems. Looking in Marvel we were hovering at around 30-40
 segments per shard back then.

 Best Regards,
 Robin

 On Friday, April 11, 2014 1:35:42 AM UTC+2, Adrien Grand wrote:

 Thanks for reporting this, the behavior is definitely unexpected. I'll
 test _optimize on very large numbers of shards to see if I can reproduce
 the issue.


  On Thu, Apr 10, 2014 at 2:10 PM, Elliott Bradshaw ebrad...@gmail.com
  wrote:

  Adrien,

 Just an FYI, after resetting the cluster, things seem to have
 improved.  Optimize calls now lead to CPU/IO activity over their 
 duration.
 Max_num_segments=1 does not seem to be working for me on any given call, 
 as
 each call would only reduce the segment count by about 600-700.  I ran 10
 calls in sequence overnight, and actually got down to 4 segments 
 (1/shard)!

 I'm glad I got the index optimized, searches are literally 10-20
 times faster without 1500/segments per shard to deal with.  It's awesome.

 That said, any thoughts on why the index wasn't merging on its own,
 or why optimize was returning prematurely?


 On Wednesday, April 9, 2014 11:10:56 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I kept the logs up over the last optimize call, and I did see an
 exception.  I Ctrl-C'd a curl optimize call before making another one, 
 but
 I don't think that that caused this exception.  The error is 
 essentially as
 follows:

 netty - Caught exception while handling client http traffic, closing
 connection [id: 0x4d8f1a90, /127.0.0.1:33480 : /127.0.0.1:9200]

 java.nio.channels.ClosedChannelException at AbstractNioWorker.
 cleanUpWriteBuffer(AbstractNioWorker.java:433)
 at AbstractNioWorker.writeFromUserCode
 at NioServerSocketPipelineSink.handleAcceptedSocket
 at NioServerSocketPipelineSink.eventSunk
 at DefaultChannelPipeline$DefaultChannelhandlerContext.sendDown
 stream
 at Channels.write
 at OneToOneEncoder.doEncode
 at 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-11 Thread Robin Wallin
Hello,

We are experiencing a related problem with 1.1.0. Segments do not seem to 
merge as they should during indexing. The optimize API does practically 
nothing in terms of lowering the segments count either. The problem 
persists through a cluster restart. The vast amount of segments seem to be 
greatly impact the performance of the cluster, in a very negative way.

We currently have 414 million documents across 3 nodes, each shard has in 
average 1200 segments(!).

With 1.0.1 we had even more documents, ~650 million, without any segment 
problems. Looking in Marvel we were hovering at around 30-40 segments per 
shard back then.

Best Regards,
Robin

On Friday, April 11, 2014 1:35:42 AM UTC+2, Adrien Grand wrote:

 Thanks for reporting this, the behavior is definitely unexpected. I'll 
 test _optimize on very large numbers of shards to see if I can reproduce 
 the issue.


 On Thu, Apr 10, 2014 at 2:10 PM, Elliott Bradshaw 
 ebrad...@gmail.comjavascript:
  wrote:

 Adrien,

 Just an FYI, after resetting the cluster, things seem to have improved.  
 Optimize calls now lead to CPU/IO activity over their duration.  
 Max_num_segments=1 does not seem to be working for me on any given call, as 
 each call would only reduce the segment count by about 600-700.  I ran 10 
 calls in sequence overnight, and actually got down to 4 segments (1/shard)!

 I'm glad I got the index optimized, searches are literally 10-20 times 
 faster without 1500/segments per shard to deal with.  It's awesome.

 That said, any thoughts on why the index wasn't merging on its own, or 
 why optimize was returning prematurely?


 On Wednesday, April 9, 2014 11:10:56 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I kept the logs up over the last optimize call, and I did see an 
 exception.  I Ctrl-C'd a curl optimize call before making another one, but 
 I don't think that that caused this exception.  The error is essentially as 
 follows:

 netty - Caught exception while handling client http traffic, closing 
 connection [id: 0x4d8f1a90, /127.0.0.1:33480 : /127.0.0.1:9200]

 java.nio.channels.ClosedChannelException at AbstractNioWorker.
 cleanUpWriteBuffer(AbstractNioWorker.java:433)
 at AbstractNioWorker.writeFromUserCode
 at NioServerSocketPipelineSink.handleAcceptedSocket
 at NioServerSocketPipelineSink.eventSunk
 at DefaultChannelPipeline$DefaultChannelhandlerContext.sendDownstream
 at Channels.write
 at OneToOneEncoder.doEncode
 at OneToOneEncoder.handleDownstream
 at DefaultChannelPipeline.sendDownstream
 at DefaultChannelPipeline.sendDownstream
 at Channels.write
 at AbstractChannel.write
 at NettyHttpChannel.sendResponse
 at RestOptimizeAction$1.onResponse(95)
 at RestOptimizeAction$1.onResponse(85)
 at TransportBroadcastOperationAction$AsyncBroadcastAction.finishHim
 at TransportBroadcastOperationAction$AsyncBroadcastAction.onOperation
 at TransportBroadcastOperationAction$AsyncBroadcastAction$2.run

 Sorry about the crappy stack trace.  Still, looks like this might point 
 to a problem!  The exception fired about an hour after I kicked off the 
 optimize.  Any thoughts?

 On Wednesday, April 9, 2014 10:06:57 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I did customize my merge policy, although I did so only because I was 
 so surprised by the number of segments left over after the load.  I'm 
 pretty sure the optimize problem was happening before I made this change, 
 but either way here are my settings:

 index : {
 merge : {
 policy : {
 max_merged_segment : 20gb,
 segments_per_tier : 5,
 floor_segment : 10mb
 },
 scheduler : concurrentmergescheduler
 }
 }

 Not sure whether this set up could be a contributing factor or not.  
 Nothing really jumps out at me in the logs.  In fact, when i kick off the 
 optimize, I don't see any logging at all.  Should I?

 I'm running the following command: curl -XPOST 
 http://localhost:9200/index/_optimize

 Thanks!


 On Wednesday, April 9, 2014 8:56:35 AM UTC-4, Adrien Grand wrote:

 Hi Elliott,

 1500 segments per shard is certainly way too much, and it is not 
 normal that optimize doesn't manage to reduce the number of segments.
  - Is there anything suspicious in the logs?
  - Have you customized the merge policy or scheduler?[1]
  - Does the issue still reproduce if you restart your cluster?

 [1] http://www.elasticsearch.org/guide/en/elasticsearch/
 reference/current/index-modules-merge.html



 On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw 
 ebrad...@gmail.comwrote:

 Any other thoughts on this?  Would 1500 segments per shard be 
 significantly impacting performance?  Have you guys noticed this 
 behavior 
 elsewhere?

 Thanks.


 On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d 
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs 
 showed cluster state updated.

 I did have to close my index prior to 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-11 Thread Nikolas Everett
For reference I'm also on 1.1.0 but I'm not seeing more segments then I
expect.  I see an average of ~28 per shard on an index I write to
constantly.  I don't write all that quickly,  50 updates a second.

Nik


On Fri, Apr 11, 2014 at 5:46 AM, Adrien Grand 
adrien.gr...@elasticsearch.com wrote:

 I managed to reproduce the issue locally, I'm looking into it.


 On Fri, Apr 11, 2014 at 9:45 AM, Robin Wallin walro...@gmail.com wrote:

 Hello,

 We are experiencing a related problem with 1.1.0. Segments do not seem to
 merge as they should during indexing. The optimize API does practically
 nothing in terms of lowering the segments count either. The problem
 persists through a cluster restart. The vast amount of segments seem to be
 greatly impact the performance of the cluster, in a very negative way.

 We currently have 414 million documents across 3 nodes, each shard has in
 average 1200 segments(!).

 With 1.0.1 we had even more documents, ~650 million, without any segment
 problems. Looking in Marvel we were hovering at around 30-40 segments per
 shard back then.

 Best Regards,
 Robin

 On Friday, April 11, 2014 1:35:42 AM UTC+2, Adrien Grand wrote:

 Thanks for reporting this, the behavior is definitely unexpected. I'll
 test _optimize on very large numbers of shards to see if I can reproduce
 the issue.


  On Thu, Apr 10, 2014 at 2:10 PM, Elliott Bradshaw 
 ebrad...@gmail.comwrote:

  Adrien,

 Just an FYI, after resetting the cluster, things seem to have
 improved.  Optimize calls now lead to CPU/IO activity over their duration.
 Max_num_segments=1 does not seem to be working for me on any given call, as
 each call would only reduce the segment count by about 600-700.  I ran 10
 calls in sequence overnight, and actually got down to 4 segments (1/shard)!

 I'm glad I got the index optimized, searches are literally 10-20 times
 faster without 1500/segments per shard to deal with.  It's awesome.

 That said, any thoughts on why the index wasn't merging on its own, or
 why optimize was returning prematurely?


 On Wednesday, April 9, 2014 11:10:56 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I kept the logs up over the last optimize call, and I did see an
 exception.  I Ctrl-C'd a curl optimize call before making another one, but
 I don't think that that caused this exception.  The error is essentially 
 as
 follows:

 netty - Caught exception while handling client http traffic, closing
 connection [id: 0x4d8f1a90, /127.0.0.1:33480 : /127.0.0.1:9200]

 java.nio.channels.ClosedChannelException at AbstractNioWorker.
 cleanUpWriteBuffer(AbstractNioWorker.java:433)
 at AbstractNioWorker.writeFromUserCode
 at NioServerSocketPipelineSink.handleAcceptedSocket
 at NioServerSocketPipelineSink.eventSunk
 at DefaultChannelPipeline$DefaultChannelhandlerContext.sendDownstream
 at Channels.write
 at OneToOneEncoder.doEncode
 at OneToOneEncoder.handleDownstream
 at DefaultChannelPipeline.sendDownstream
 at DefaultChannelPipeline.sendDownstream
 at Channels.write
 at AbstractChannel.write
 at NettyHttpChannel.sendResponse
 at RestOptimizeAction$1.onResponse(95)
 at RestOptimizeAction$1.onResponse(85)
 at TransportBroadcastOperationAction$AsyncBroadcastAction.finishHim
 at TransportBroadcastOperationAction$AsyncBroadcastAction.onOperation
 at TransportBroadcastOperationAction$AsyncBroadcastAction$2.run

 Sorry about the crappy stack trace.  Still, looks like this might
 point to a problem!  The exception fired about an hour after I kicked off
 the optimize.  Any thoughts?

 On Wednesday, April 9, 2014 10:06:57 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I did customize my merge policy, although I did so only because I was
 so surprised by the number of segments left over after the load.  I'm
 pretty sure the optimize problem was happening before I made this change,
 but either way here are my settings:

 index : {
 merge : {
 policy : {
 max_merged_segment : 20gb,
 segments_per_tier : 5,
 floor_segment : 10mb
 },
 scheduler : concurrentmergescheduler
 }
 }

 Not sure whether this set up could be a contributing factor or not.
 Nothing really jumps out at me in the logs.  In fact, when i kick off the
 optimize, I don't see any logging at all.  Should I?

 I'm running the following command: curl -XPOST
 http://localhost:9200/index/_optimize

 Thanks!


 On Wednesday, April 9, 2014 8:56:35 AM UTC-4, Adrien Grand wrote:

 Hi Elliott,

 1500 segments per shard is certainly way too much, and it is not
 normal that optimize doesn't manage to reduce the number of segments.
  - Is there anything suspicious in the logs?
  - Have you customized the merge policy or scheduler?[1]
  - Does the issue still reproduce if you restart your cluster?

 [1] http://www.elasticsearch.org/guide/en/elasticsearch/referenc
 e/current/index-modules-merge.html



 On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw ebrad...@gmail.com
  wrote:

 Any other thoughts on this?  Would 1500 segments per shard be
 significantly impacting performance? 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-10 Thread Elliott Bradshaw
Adrien,

Just an FYI, after resetting the cluster, things seem to have improved.  
Optimize calls now lead to CPU/IO activity over their duration.  
Max_num_segments=1 does not seem to be working for me on any given call, as 
each call would only reduce the segment count by about 600-700.  I ran 10 
calls in sequence overnight, and actually got down to 4 segments (1/shard)!

I'm glad I got the index optimized, searches are literally 10-20 times 
faster without 1500/segments per shard to deal with.  It's awesome.

That said, any thoughts on why the index wasn't merging on its own, or why 
optimize was returning prematurely?

On Wednesday, April 9, 2014 11:10:56 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I kept the logs up over the last optimize call, and I did see an 
 exception.  I Ctrl-C'd a curl optimize call before making another one, but 
 I don't think that that caused this exception.  The error is essentially as 
 follows:

 netty - Caught exception while handling client http traffic, closing 
 connection [id: 0x4d8f1a90, /127.0.0.1:33480 : /127.0.0.1:9200]

 java.nio.channels.ClosedChannelException at 
 AbstractNioWorker.cleanUpWriteBuffer(AbstractNioWorker.java:433)
 at AbstractNioWorker.writeFromUserCode
 at NioServerSocketPipelineSink.handleAcceptedSocket
 at NioServerSocketPipelineSink.eventSunk
 at DefaultChannelPipeline$DefaultChannelhandlerContext.sendDownstream
 at Channels.write
 at OneToOneEncoder.doEncode
 at OneToOneEncoder.handleDownstream
 at DefaultChannelPipeline.sendDownstream
 at DefaultChannelPipeline.sendDownstream
 at Channels.write
 at AbstractChannel.write
 at NettyHttpChannel.sendResponse
 at RestOptimizeAction$1.onResponse(95)
 at RestOptimizeAction$1.onResponse(85)
 at TransportBroadcastOperationAction$AsyncBroadcastAction.finishHim
 at TransportBroadcastOperationAction$AsyncBroadcastAction.onOperation
 at TransportBroadcastOperationAction$AsyncBroadcastAction$2.run

 Sorry about the crappy stack trace.  Still, looks like this might point to 
 a problem!  The exception fired about an hour after I kicked off the 
 optimize.  Any thoughts?

 On Wednesday, April 9, 2014 10:06:57 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I did customize my merge policy, although I did so only because I was so 
 surprised by the number of segments left over after the load.  I'm pretty 
 sure the optimize problem was happening before I made this change, but 
 either way here are my settings:

 index : {
 merge : {
 policy : {
 max_merged_segment : 20gb,
 segments_per_tier : 5,
 floor_segment : 10mb
 },
 scheduler : concurrentmergescheduler
 }
 }

 Not sure whether this set up could be a contributing factor or not.  
 Nothing really jumps out at me in the logs.  In fact, when i kick off the 
 optimize, I don't see any logging at all.  Should I?

 I'm running the following command: curl -XPOST 
 http://localhost:9200/index/_optimize

 Thanks!


 On Wednesday, April 9, 2014 8:56:35 AM UTC-4, Adrien Grand wrote:

 Hi Elliott,

 1500 segments per shard is certainly way too much, and it is not normal 
 that optimize doesn't manage to reduce the number of segments.
  - Is there anything suspicious in the logs?
  - Have you customized the merge policy or scheduler?[1]
  - Does the issue still reproduce if you restart your cluster?

 [1] 
 http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-merge.html



 On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw ebrad...@gmail.comwrote:

 Any other thoughts on this?  Would 1500 segments per shard be 
 significantly impacting performance?  Have you guys noticed this behavior 
 elsewhere?

 Thanks.


 On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d 
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs showed 
 cluster state updated.

 I did have to close my index prior to changing the setting and reopen 
 afterward.


 I've since began another optimize, but again it doesn't look like much 
 is happening.  The optimize isn't returning and the total CPU usage on 
 every node is holding at about 2% of a single core.  I would copy a 
 hot_threads stack trace, but I'm unfortunately on a closed network and 
 this 
 isn't possible.  I can tell you that refreshes of hot_threads show vary 
 little happening.  The occasional [merge] thread (always in a 
 LinkedTransferQueue.awaitMatch() state) or [optimize] (doing nothing 
 on a waitForMerge() call) thread shows up, but it's always consuming 0-1% 
 CPU.  It sure feels like something isn't right.  Any thoughts?




 On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand adrien...@elasticsearch.
 com wrote:

 Did you see a message in the logs confirming that the setting has 
 been updated? It would be interesting to see the output of hot 
 threads[1] 
 to see what your node is doing.

 [1] 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-10 Thread Adrien Grand
Thanks for reporting this, the behavior is definitely unexpected. I'll test
_optimize on very large numbers of shards to see if I can reproduce the
issue.


On Thu, Apr 10, 2014 at 2:10 PM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Adrien,

 Just an FYI, after resetting the cluster, things seem to have improved.
 Optimize calls now lead to CPU/IO activity over their duration.
 Max_num_segments=1 does not seem to be working for me on any given call, as
 each call would only reduce the segment count by about 600-700.  I ran 10
 calls in sequence overnight, and actually got down to 4 segments (1/shard)!

 I'm glad I got the index optimized, searches are literally 10-20 times
 faster without 1500/segments per shard to deal with.  It's awesome.

 That said, any thoughts on why the index wasn't merging on its own, or why
 optimize was returning prematurely?


 On Wednesday, April 9, 2014 11:10:56 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I kept the logs up over the last optimize call, and I did see an
 exception.  I Ctrl-C'd a curl optimize call before making another one, but
 I don't think that that caused this exception.  The error is essentially as
 follows:

 netty - Caught exception while handling client http traffic, closing
 connection [id: 0x4d8f1a90, /127.0.0.1:33480 : /127.0.0.1:9200]

 java.nio.channels.ClosedChannelException at AbstractNioWorker.
 cleanUpWriteBuffer(AbstractNioWorker.java:433)
 at AbstractNioWorker.writeFromUserCode
 at NioServerSocketPipelineSink.handleAcceptedSocket
 at NioServerSocketPipelineSink.eventSunk
 at DefaultChannelPipeline$DefaultChannelhandlerContext.sendDownstream
 at Channels.write
 at OneToOneEncoder.doEncode
 at OneToOneEncoder.handleDownstream
 at DefaultChannelPipeline.sendDownstream
 at DefaultChannelPipeline.sendDownstream
 at Channels.write
 at AbstractChannel.write
 at NettyHttpChannel.sendResponse
 at RestOptimizeAction$1.onResponse(95)
 at RestOptimizeAction$1.onResponse(85)
 at TransportBroadcastOperationAction$AsyncBroadcastAction.finishHim
 at TransportBroadcastOperationAction$AsyncBroadcastAction.onOperation
 at TransportBroadcastOperationAction$AsyncBroadcastAction$2.run

 Sorry about the crappy stack trace.  Still, looks like this might point
 to a problem!  The exception fired about an hour after I kicked off the
 optimize.  Any thoughts?

 On Wednesday, April 9, 2014 10:06:57 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I did customize my merge policy, although I did so only because I was so
 surprised by the number of segments left over after the load.  I'm pretty
 sure the optimize problem was happening before I made this change, but
 either way here are my settings:

 index : {
 merge : {
 policy : {
 max_merged_segment : 20gb,
 segments_per_tier : 5,
 floor_segment : 10mb
 },
 scheduler : concurrentmergescheduler
 }
 }

 Not sure whether this set up could be a contributing factor or not.
 Nothing really jumps out at me in the logs.  In fact, when i kick off the
 optimize, I don't see any logging at all.  Should I?

 I'm running the following command: curl -XPOST
 http://localhost:9200/index/_optimize

 Thanks!


 On Wednesday, April 9, 2014 8:56:35 AM UTC-4, Adrien Grand wrote:

 Hi Elliott,

 1500 segments per shard is certainly way too much, and it is not normal
 that optimize doesn't manage to reduce the number of segments.
  - Is there anything suspicious in the logs?
  - Have you customized the merge policy or scheduler?[1]
  - Does the issue still reproduce if you restart your cluster?

 [1] http://www.elasticsearch.org/guide/en/elasticsearch/
 reference/current/index-modules-merge.html



 On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw ebrad...@gmail.comwrote:

 Any other thoughts on this?  Would 1500 segments per shard be
 significantly impacting performance?  Have you guys noticed this behavior
 elsewhere?

 Thanks.


 On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs showed
 cluster state updated.

 I did have to close my index prior to changing the setting and reopen
 afterward.


 I've since began another optimize, but again it doesn't look like
 much is happening.  The optimize isn't returning and the total CPU usage 
 on
 every node is holding at about 2% of a single core.  I would copy a
 hot_threads stack trace, but I'm unfortunately on a closed network and 
 this
 isn't possible.  I can tell you that refreshes of hot_threads show vary
 little happening.  The occasional [merge] thread (always in a
 LinkedTransferQueue.awaitMatch() state) or [optimize] (doing nothing
 on a waitForMerge() call) thread shows up, but it's always consuming 0-1%
 CPU.  It sure feels like something isn't right.  Any thoughts?




 On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand 
 adrien...@elasticsearch.com wrote:

 Did you 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-09 Thread Elliott Bradshaw
Any other thoughts on this?  Would 1500 segments per shard be significantly 
impacting performance?  Have you guys noticed this behavior elsewhere?

Thanks.

On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d 
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs showed 
 cluster state updated.

 I did have to close my index prior to changing the setting and reopen 
 afterward.


 I've since began another optimize, but again it doesn't look like much is 
 happening.  The optimize isn't returning and the total CPU usage on every 
 node is holding at about 2% of a single core.  I would copy a hot_threads 
 stack trace, but I'm unfortunately on a closed network and this isn't 
 possible.  I can tell you that refreshes of hot_threads show vary little 
 happening.  The occasional [merge] thread (always in a 
 LinkedTransferQueue.awaitMatch() state) or [optimize] (doing nothing on a 
 waitForMerge() call) thread shows up, but it's always consuming 0-1% CPU.  
 It sure feels like something isn't right.  Any thoughts?




 On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand 
 adrien.gr...@elasticsearch.com wrote:

 Did you see a message in the logs confirming that the setting has been 
 updated? It would be interesting to see the output of hot threads[1] to see 
 what your node is doing.

 [1] 
 http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/cluster-nodes-hot-threads.html


 On Fri, Apr 4, 2014 at 7:18 PM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Yes. I have run max_num_segments=1 every time.


 On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
 michael.s...@serenesoftware.com wrote:

 Have you tried max_num_segments=1 on your optimize? 

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw ebradsh...@gmail.com
  wrote:

 Any thoughts on this?  I've run optimize several more times, and the 
 number of segments falls each time, but I'm still over 1000 segments per 
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening 
 in the background, but I'm still seeing over 6500 segments.  Even after 
 setting max_num_segments=5.  Does this seem right?  Queries are a little 
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair 
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node 
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on 
 each.  I've batch loaded approximately 86 million documents into a 
 single 
 index (4 shards) and have started benchmarking cross_field/multi_match 
 queries on them.  The index has one replica and takes up a total of 
 111G.  
 I've run several batches of warming queries, but queries are not as 
 fast as 
 I had hoped, approximately 400-500ms each.  Given that *top *(on 
 Centos) shows 5-8 GB of free memory on each server, I would assume that 
 the 
 entire index has been paged into memory (I had worried about disk 
 performance previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is 
 composed of  7000 segments.  This seemed high to me, but maybe it's 
 appropriate.  Regardless, I dispatched an optimize command, but I am 
 not 
 seeing any progress and the command has not returned.  Current merges 
 remains at zero, and the segment count is not changing.  Checking out 
 hot 
 threads in ElasticHQ, I initially saw an optimize call in the stack 
 that 
 was blocked on a waitForMerge call.  This however has disappeared, and 
 I'm 
 seeing no evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else 
 had similar issues.  This is the second time I have tried to optimize 
 an 
 index since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  -- 
 You received this message because you are subscribed to the Google 
 Groups elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send 
 an email to elasticsearch+unsubscr...@googlegroups.com.
  To view this discussion on the web visit 
 https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.comhttps://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


  -- 
 You received this message because you are subscribed to a topic in the 
 Google Groups elasticsearch group.
 To unsubscribe from this topic, visit 
 https://groups.google.com/d/topic/elasticsearch/kqTRRADQBwc/unsubscribe
 .
 To 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-09 Thread Adrien Grand
Hi Elliott,

1500 segments per shard is certainly way too much, and it is not normal
that optimize doesn't manage to reduce the number of segments.
 - Is there anything suspicious in the logs?
 - Have you customized the merge policy or scheduler?[1]
 - Does the issue still reproduce if you restart your cluster?

[1]
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-merge.html



On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Any other thoughts on this?  Would 1500 segments per shard be
 significantly impacting performance?  Have you guys noticed this behavior
 elsewhere?

 Thanks.


 On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs showed
 cluster state updated.

 I did have to close my index prior to changing the setting and reopen
 afterward.


 I've since began another optimize, but again it doesn't look like much is
 happening.  The optimize isn't returning and the total CPU usage on every
 node is holding at about 2% of a single core.  I would copy a hot_threads
 stack trace, but I'm unfortunately on a closed network and this isn't
 possible.  I can tell you that refreshes of hot_threads show vary little
 happening.  The occasional [merge] thread (always in a 
 LinkedTransferQueue.awaitMatch()
 state) or [optimize] (doing nothing on a waitForMerge() call) thread shows
 up, but it's always consuming 0-1% CPU.  It sure feels like something isn't
 right.  Any thoughts?




 On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand adrien.grand@elasticsearch.
 com wrote:

 Did you see a message in the logs confirming that the setting has been
 updated? It would be interesting to see the output of hot threads[1] to see
 what your node is doing.

 [1] http://www.elasticsearch.org/guide/en/elasticsearch/
 reference/current/cluster-nodes-hot-threads.html


 On Fri, Apr 4, 2014 at 7:18 PM, Elliott Bradshaw 
 ebradsh...@gmail.comwrote:

 Yes. I have run max_num_segments=1 every time.


 On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
 michael.s...@serenesoftware.com wrote:

 Have you tried max_num_segments=1 on your optimize?

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw 
 ebradsh...@gmail.com wrote:

 Any thoughts on this?  I've run optimize several more times, and the
 number of segments falls each time, but I'm still over 1000 segments per
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening
 in the background, but I'm still seeing over 6500 segments.  Even after
 setting max_num_segments=5.  Does this seem right?  Queries are a little
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on
 each.  I've batch loaded approximately 86 million documents into a 
 single
 index (4 shards) and have started benchmarking cross_field/multi_match
 queries on them.  The index has one replica and takes up a total of 
 111G.
 I've run several batches of warming queries, but queries are not as 
 fast as
 I had hoped, approximately 400-500ms each.  Given that *top *(on
 Centos) shows 5-8 GB of free memory on each server, I would assume 
 that the
 entire index has been paged into memory (I had worried about disk
 performance previously, as we are working in a virtualized 
 environment).

 A stats query on the index in questions shows that the index is
 composed of  7000 segments.  This seemed high to me, but maybe it's
 appropriate.  Regardless, I dispatched an optimize command, but I am 
 not
 seeing any progress and the command has not returned.  Current merges
 remains at zero, and the segment count is not changing.  Checking out 
 hot
 threads in ElasticHQ, I initially saw an optimize call in the stack 
 that
 was blocked on a waitForMerge call.  This however has disappeared, and 
 I'm
 seeing no evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else
 had similar issues.  This is the second time I have tried to optimize 
 an
 index since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  --
 You received this message because you are subscribed to the Google
 Groups elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it,
 send an email to elasticsearch+unsubscr...@googlegroups.com.
  To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/5391291f-
 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-09 Thread Elliott Bradshaw
Hi Adrien,

I did customize my merge policy, although I did so only because I was so 
surprised by the number of segments left over after the load.  I'm pretty 
sure the optimize problem was happening before I made this change, but 
either way here are my settings:

index : {
merge : {
policy : {
max_merged_segment : 20gb,
segments_per_tier : 5,
floor_segment : 10mb
},
scheduler : concurrentmergescheduler
}
}

Not sure whether this set up could be a contributing factor or not.  
Nothing really jumps out at me in the logs.  In fact, when i kick off the 
optimize, I don't see any logging at all.  Should I?

I'm running the following command: curl -XPOST 
http://localhost:9200/index/_optimize

Thanks!


On Wednesday, April 9, 2014 8:56:35 AM UTC-4, Adrien Grand wrote:

 Hi Elliott,

 1500 segments per shard is certainly way too much, and it is not normal 
 that optimize doesn't manage to reduce the number of segments.
  - Is there anything suspicious in the logs?
  - Have you customized the merge policy or scheduler?[1]
  - Does the issue still reproduce if you restart your cluster?

 [1] 
 http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-merge.html



 On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw 
 ebrad...@gmail.comjavascript:
  wrote:

 Any other thoughts on this?  Would 1500 segments per shard be 
 significantly impacting performance?  Have you guys noticed this behavior 
 elsewhere?

 Thanks.


 On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d 
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs showed 
 cluster state updated.

 I did have to close my index prior to changing the setting and reopen 
 afterward.


 I've since began another optimize, but again it doesn't look like much 
 is happening.  The optimize isn't returning and the total CPU usage on 
 every node is holding at about 2% of a single core.  I would copy a 
 hot_threads stack trace, but I'm unfortunately on a closed network and this 
 isn't possible.  I can tell you that refreshes of hot_threads show vary 
 little happening.  The occasional [merge] thread (always in a 
 LinkedTransferQueue.awaitMatch() state) or [optimize] (doing nothing on 
 a waitForMerge() call) thread shows up, but it's always consuming 0-1% 
 CPU.  It sure feels like something isn't right.  Any thoughts?




 On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand adrien...@elasticsearch.
 com javascript: wrote:

 Did you see a message in the logs confirming that the setting has been 
 updated? It would be interesting to see the output of hot threads[1] to 
 see 
 what your node is doing.

 [1] http://www.elasticsearch.org/guide/en/elasticsearch/
 reference/current/cluster-nodes-hot-threads.html


 On Fri, Apr 4, 2014 at 7:18 PM, Elliott Bradshaw 
 ebrad...@gmail.comjavascript:
  wrote:

 Yes. I have run max_num_segments=1 every time.


 On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
 michae...@serenesoftware.com javascript: wrote:

 Have you tried max_num_segments=1 on your optimize? 

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw 
 ebrad...@gmail.comjavascript:
  wrote:

 Any thoughts on this?  I've run optimize several more times, and the 
 number of segments falls each time, but I'm still over 1000 segments 
 per 
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was 
 happening in the background, but I'm still seeing over 6500 segments.  
 Even 
 after setting max_num_segments=5.  Does this seem right?  Queries are 
 a 
 little faster (350-400ms) but still not great.  Bigdesk is still 
 showing a 
 fair amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node 
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on 
 each.  I've batch loaded approximately 86 million documents into a 
 single 
 index (4 shards) and have started benchmarking 
 cross_field/multi_match 
 queries on them.  The index has one replica and takes up a total of 
 111G.  
 I've run several batches of warming queries, but queries are not as 
 fast as 
 I had hoped, approximately 400-500ms each.  Given that *top *(on 
 Centos) shows 5-8 GB of free memory on each server, I would assume 
 that the 
 entire index has been paged into memory (I had worried about disk 
 performance previously, as we are working in a virtualized 
 environment).

 A stats query on the index in questions shows that the index is 
 composed of  7000 segments.  This seemed high to me, but maybe it's 
 appropriate.  Regardless, I dispatched an optimize command, but I am 
 not 
 seeing any progress and the command has not returned.  Current merges 
 remains at zero, 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-09 Thread Elliott Bradshaw
Hi Adrien,

I kept the logs up over the last optimize call, and I did see an 
exception.  I Ctrl-C'd a curl optimize call before making another one, but 
I don't think that that caused this exception.  The error is essentially as 
follows:

netty - Caught exception while handling client http traffic, closing 
connection [id: 0x4d8f1a90, /127.0.0.1:33480 : /127.0.0.1:9200]

java.nio.channels.ClosedChannelException at 
AbstractNioWorker.cleanUpWriteBuffer(AbstractNioWorker.java:433)
at AbstractNioWorker.writeFromUserCode
at NioServerSocketPipelineSink.handleAcceptedSocket
at NioServerSocketPipelineSink.eventSunk
at DefaultChannelPipeline$DefaultChannelhandlerContext.sendDownstream
at Channels.write
at OneToOneEncoder.doEncode
at OneToOneEncoder.handleDownstream
at DefaultChannelPipeline.sendDownstream
at DefaultChannelPipeline.sendDownstream
at Channels.write
at AbstractChannel.write
at NettyHttpChannel.sendResponse
at RestOptimizeAction$1.onResponse(95)
at RestOptimizeAction$1.onResponse(85)
at TransportBroadcastOperationAction$AsyncBroadcastAction.finishHim
at TransportBroadcastOperationAction$AsyncBroadcastAction.onOperation
at TransportBroadcastOperationAction$AsyncBroadcastAction$2.run

Sorry about the crappy stack trace.  Still, looks like this might point to 
a problem!  The exception fired about an hour after I kicked off the 
optimize.  Any thoughts?

On Wednesday, April 9, 2014 10:06:57 AM UTC-4, Elliott Bradshaw wrote:

 Hi Adrien,

 I did customize my merge policy, although I did so only because I was so 
 surprised by the number of segments left over after the load.  I'm pretty 
 sure the optimize problem was happening before I made this change, but 
 either way here are my settings:

 index : {
 merge : {
 policy : {
 max_merged_segment : 20gb,
 segments_per_tier : 5,
 floor_segment : 10mb
 },
 scheduler : concurrentmergescheduler
 }
 }

 Not sure whether this set up could be a contributing factor or not.  
 Nothing really jumps out at me in the logs.  In fact, when i kick off the 
 optimize, I don't see any logging at all.  Should I?

 I'm running the following command: curl -XPOST 
 http://localhost:9200/index/_optimize

 Thanks!


 On Wednesday, April 9, 2014 8:56:35 AM UTC-4, Adrien Grand wrote:

 Hi Elliott,

 1500 segments per shard is certainly way too much, and it is not normal 
 that optimize doesn't manage to reduce the number of segments.
  - Is there anything suspicious in the logs?
  - Have you customized the merge policy or scheduler?[1]
  - Does the issue still reproduce if you restart your cluster?

 [1] 
 http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-merge.html



 On Wed, Apr 9, 2014 at 2:38 PM, Elliott Bradshaw ebrad...@gmail.comwrote:

 Any other thoughts on this?  Would 1500 segments per shard be 
 significantly impacting performance?  Have you guys noticed this behavior 
 elsewhere?

 Thanks.


 On Monday, April 7, 2014 8:56:38 AM UTC-4, Elliott Bradshaw wrote:

 Adrian,

 I ran the following command:

 curl -XPUT http://localhost:9200/_settings -d 
 '{indices.store.throttle.max_bytes_per_sec : 10gb}'

 and received a { acknowledged : true } response.  The logs showed 
 cluster state updated.

 I did have to close my index prior to changing the setting and reopen 
 afterward.


 I've since began another optimize, but again it doesn't look like much 
 is happening.  The optimize isn't returning and the total CPU usage on 
 every node is holding at about 2% of a single core.  I would copy a 
 hot_threads stack trace, but I'm unfortunately on a closed network and 
 this 
 isn't possible.  I can tell you that refreshes of hot_threads show vary 
 little happening.  The occasional [merge] thread (always in a 
 LinkedTransferQueue.awaitMatch() state) or [optimize] (doing nothing 
 on a waitForMerge() call) thread shows up, but it's always consuming 0-1% 
 CPU.  It sure feels like something isn't right.  Any thoughts?




 On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand adrien...@elasticsearch.
 com wrote:

 Did you see a message in the logs confirming that the setting has been 
 updated? It would be interesting to see the output of hot threads[1] to 
 see 
 what your node is doing.

 [1] http://www.elasticsearch.org/guide/en/elasticsearch/
 reference/current/cluster-nodes-hot-threads.html


 On Fri, Apr 4, 2014 at 7:18 PM, Elliott Bradshaw 
 ebrad...@gmail.comwrote:

 Yes. I have run max_num_segments=1 every time.


 On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
 michae...@serenesoftware.com wrote:

 Have you tried max_num_segments=1 on your optimize? 

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw 
 ebrad...@gmail.com wrote:

 Any thoughts on this?  I've run optimize several more times, and 
 the number of segments falls each time, but I'm still over 1000 
 segments 
 per shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw 
 wrote:

 OK.  Optimize finally returned, so I 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-09 Thread joergpra...@gmail.com
The exception is just a side effect because you pressed ctrl-c and the
response could not be transmitted back, it does not point to the problem.

You should use

http://localhost:9200/index/_optimize?max_num_segments=1

instead of

http://localhost:9200/index/_optimize

Jörg

-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAKdsXoGFJw%3D7MDw8EW0DtnpqzVturkb9UmdyiiZPUJ8m-3HaMg%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-09 Thread Elliott Bradshaw
Thanks Jorg.  That makes sense.  I am actually using max_num_segments=1,
just forgot to add it...


On Wed, Apr 9, 2014 at 11:20 AM, joergpra...@gmail.com 
joergpra...@gmail.com wrote:

 The exception is just a side effect because you pressed ctrl-c and the
 response could not be transmitted back, it does not point to the problem.

 You should use

 http://localhost:9200/index/_optimize?max_num_segments=1

 instead of

 http://localhost:9200/index/_optimize

 Jörg

  --
 You received this message because you are subscribed to a topic in the
 Google Groups elasticsearch group.
 To unsubscribe from this topic, visit
 https://groups.google.com/d/topic/elasticsearch/kqTRRADQBwc/unsubscribe.
 To unsubscribe from this group and all its topics, send an email to
 elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/CAKdsXoGFJw%3D7MDw8EW0DtnpqzVturkb9UmdyiiZPUJ8m-3HaMg%40mail.gmail.comhttps://groups.google.com/d/msgid/elasticsearch/CAKdsXoGFJw%3D7MDw8EW0DtnpqzVturkb9UmdyiiZPUJ8m-3HaMg%40mail.gmail.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAGCt%2BFuGLJaBLfB1PyFWKRS8HBJ81%2BwN1BFua6Ua5H7QyeVybQ%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-07 Thread Elliott Bradshaw
Adrian,

I ran the following command:

curl -XPUT http://localhost:9200/_settings -d
'{indices.store.throttle.max_bytes_per_sec : 10gb}'

and received a { acknowledged : true } response.  The logs showed
cluster state updated.

I did have to close my index prior to changing the setting and reopen
afterward.


I've since began another optimize, but again it doesn't look like much is
happening.  The optimize isn't returning and the total CPU usage on every
node is holding at about 2% of a single core.  I would copy a hot_threads
stack trace, but I'm unfortunately on a closed network and this isn't
possible.  I can tell you that refreshes of hot_threads show vary little
happening.  The occasional [merge] thread (always in a
LinkedTransferQueue.awaitMatch() state) or [optimize] (doing nothing on a
waitForMerge() call) thread shows up, but it's always consuming 0-1% CPU.
It sure feels like something isn't right.  Any thoughts?




On Fri, Apr 4, 2014 at 3:24 PM, Adrien Grand adrien.gr...@elasticsearch.com
 wrote:

 Did you see a message in the logs confirming that the setting has been
 updated? It would be interesting to see the output of hot threads[1] to see
 what your node is doing.

 [1]
 http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/cluster-nodes-hot-threads.html


 On Fri, Apr 4, 2014 at 7:18 PM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Yes. I have run max_num_segments=1 every time.


 On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
 michael.s...@serenesoftware.com wrote:

 Have you tried max_num_segments=1 on your optimize?

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw 
 ebradsh...@gmail.comwrote:

 Any thoughts on this?  I've run optimize several more times, and the
 number of segments falls each time, but I'm still over 1000 segments per
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening
 in the background, but I'm still seeing over 6500 segments.  Even after
 setting max_num_segments=5.  Does this seem right?  Queries are a little
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on
 each.  I've batch loaded approximately 86 million documents into a single
 index (4 shards) and have started benchmarking cross_field/multi_match
 queries on them.  The index has one replica and takes up a total of 111G.
 I've run several batches of warming queries, but queries are not as fast 
 as
 I had hoped, approximately 400-500ms each.  Given that *top *(on
 Centos) shows 5-8 GB of free memory on each server, I would assume that 
 the
 entire index has been paged into memory (I had worried about disk
 performance previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is
 composed of  7000 segments.  This seemed high to me, but maybe it's
 appropriate.  Regardless, I dispatched an optimize command, but I am not
 seeing any progress and the command has not returned.  Current merges
 remains at zero, and the segment count is not changing.  Checking out hot
 threads in ElasticHQ, I initially saw an optimize call in the stack that
 was blocked on a waitForMerge call.  This however has disappeared, and 
 I'm
 seeing no evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else
 had similar issues.  This is the second time I have tried to optimize an
 index since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  --
 You received this message because you are subscribed to the Google
 Groups elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send
 an email to elasticsearch+unsubscr...@googlegroups.com.
  To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.comhttps://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


  --
 You received this message because you are subscribed to a topic in the
 Google Groups elasticsearch group.
 To unsubscribe from this topic, visit
 https://groups.google.com/d/topic/elasticsearch/kqTRRADQBwc/unsubscribe.
 To unsubscribe from this group and all its topics, send an email to
 elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 

Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-04 Thread Elliott Bradshaw
Any thoughts on this?  I've run optimize several more times, and the number 
of segments falls each time, but I'm still over 1000 segments per shard.  
Has anyone else run into something similar?

On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening in 
 the background, but I'm still seeing over 6500 segments.  Even after 
 setting max_num_segments=5.  Does this seem right?  Queries are a little 
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair 
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node 
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on 
 each.  I've batch loaded approximately 86 million documents into a single 
 index (4 shards) and have started benchmarking cross_field/multi_match 
 queries on them.  The index has one replica and takes up a total of 111G.  
 I've run several batches of warming queries, but queries are not as fast as 
 I had hoped, approximately 400-500ms each.  Given that *top *(on Centos) 
 shows 5-8 GB of free memory on each server, I would assume that the entire 
 index has been paged into memory (I had worried about disk performance 
 previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is composed 
 of  7000 segments.  This seemed high to me, but maybe it's appropriate.  
 Regardless, I dispatched an optimize command, but I am not seeing any 
 progress and the command has not returned.  Current merges remains at zero, 
 and the segment count is not changing.  Checking out hot threads in 
 ElasticHQ, I initially saw an optimize call in the stack that was blocked 
 on a waitForMerge call.  This however has disappeared, and I'm seeing no 
 evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else had 
 similar issues.  This is the second time I have tried to optimize an index 
 since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott



-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-04 Thread Adrien Grand
Elasticsearch throttles merges by default so that they don't slow search
down too much. This is usually preferable for read/writes loads, but in
your case it looks like you batch-indexed a lot of documents at once and
merges couldn't keep up with the indexing rate so you ended up with a very
high number of segments. The thing is that merge throttling also applies to
optimize calls, which might explain why your calls to the optimize API last
forever.

Could you try to disable merge throttling[1] before running a call to
optimize again to see if the situation improves?

[1]
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/index-modules-store.html#store-throttling


On Fri, Apr 4, 2014 at 5:27 PM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Any thoughts on this?  I've run optimize several more times, and the
 number of segments falls each time, but I'm still over 1000 segments per
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening in
 the background, but I'm still seeing over 6500 segments.  Even after
 setting max_num_segments=5.  Does this seem right?  Queries are a little
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on
 each.  I've batch loaded approximately 86 million documents into a single
 index (4 shards) and have started benchmarking cross_field/multi_match
 queries on them.  The index has one replica and takes up a total of 111G.
 I've run several batches of warming queries, but queries are not as fast as
 I had hoped, approximately 400-500ms each.  Given that *top *(on
 Centos) shows 5-8 GB of free memory on each server, I would assume that the
 entire index has been paged into memory (I had worried about disk
 performance previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is composed
 of  7000 segments.  This seemed high to me, but maybe it's appropriate.
 Regardless, I dispatched an optimize command, but I am not seeing any
 progress and the command has not returned.  Current merges remains at zero,
 and the segment count is not changing.  Checking out hot threads in
 ElasticHQ, I initially saw an optimize call in the stack that was blocked
 on a waitForMerge call.  This however has disappeared, and I'm seeing no
 evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else had
 similar issues.  This is the second time I have tried to optimize an index
 since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  --
 You received this message because you are subscribed to the Google Groups
 elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send an
 email to elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.comhttps://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.




-- 
Adrien Grand

-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAL6Z4j6WMRx8x-rJJi3KS2CZUu9wSbX8Vmuy48CpHFM_jUCXdw%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-04 Thread Michael Sick
Have you tried max_num_segments=1 on your optimize?

On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Any thoughts on this?  I've run optimize several more times, and the
 number of segments falls each time, but I'm still over 1000 segments per
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening in
 the background, but I'm still seeing over 6500 segments.  Even after
 setting max_num_segments=5.  Does this seem right?  Queries are a little
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on
 each.  I've batch loaded approximately 86 million documents into a single
 index (4 shards) and have started benchmarking cross_field/multi_match
 queries on them.  The index has one replica and takes up a total of 111G.
 I've run several batches of warming queries, but queries are not as fast as
 I had hoped, approximately 400-500ms each.  Given that *top *(on
 Centos) shows 5-8 GB of free memory on each server, I would assume that the
 entire index has been paged into memory (I had worried about disk
 performance previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is composed
 of  7000 segments.  This seemed high to me, but maybe it's appropriate.
 Regardless, I dispatched an optimize command, but I am not seeing any
 progress and the command has not returned.  Current merges remains at zero,
 and the segment count is not changing.  Checking out hot threads in
 ElasticHQ, I initially saw an optimize call in the stack that was blocked
 on a waitForMerge call.  This however has disappeared, and I'm seeing no
 evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else had
 similar issues.  This is the second time I have tried to optimize an index
 since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  --
 You received this message because you are subscribed to the Google Groups
 elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send an
 email to elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.comhttps://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAP8axnD7BUziGct2%3Db%3DfupaKYFnA5fR2TBsxHoURJumHSyODFA%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-04 Thread Elliott Bradshaw
Yes. I have run max_num_segments=1 every time.


On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
michael.s...@serenesoftware.com wrote:

 Have you tried max_num_segments=1 on your optimize?

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Any thoughts on this?  I've run optimize several more times, and the
 number of segments falls each time, but I'm still over 1000 segments per
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening in
 the background, but I'm still seeing over 6500 segments.  Even after
 setting max_num_segments=5.  Does this seem right?  Queries are a little
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on
 each.  I've batch loaded approximately 86 million documents into a single
 index (4 shards) and have started benchmarking cross_field/multi_match
 queries on them.  The index has one replica and takes up a total of 111G.
 I've run several batches of warming queries, but queries are not as fast as
 I had hoped, approximately 400-500ms each.  Given that *top *(on
 Centos) shows 5-8 GB of free memory on each server, I would assume that the
 entire index has been paged into memory (I had worried about disk
 performance previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is
 composed of  7000 segments.  This seemed high to me, but maybe it's
 appropriate.  Regardless, I dispatched an optimize command, but I am not
 seeing any progress and the command has not returned.  Current merges
 remains at zero, and the segment count is not changing.  Checking out hot
 threads in ElasticHQ, I initially saw an optimize call in the stack that
 was blocked on a waitForMerge call.  This however has disappeared, and I'm
 seeing no evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else had
 similar issues.  This is the second time I have tried to optimize an index
 since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  --
 You received this message because you are subscribed to the Google Groups
 elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send an
 email to elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.comhttps://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


  --
 You received this message because you are subscribed to a topic in the
 Google Groups elasticsearch group.
 To unsubscribe from this topic, visit
 https://groups.google.com/d/topic/elasticsearch/kqTRRADQBwc/unsubscribe.
 To unsubscribe from this group and all its topics, send an email to
 elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/CAP8axnD7BUziGct2%3Db%3DfupaKYFnA5fR2TBsxHoURJumHSyODFA%40mail.gmail.comhttps://groups.google.com/d/msgid/elasticsearch/CAP8axnD7BUziGct2%3Db%3DfupaKYFnA5fR2TBsxHoURJumHSyODFA%40mail.gmail.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAGCt%2BFvoSQTvv%2B6G%3D3GOX27AuYdEwLiW%3Demc0JTouT9%2BBeUk_A%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-04 Thread Adrien Grand
Did you see a message in the logs confirming that the setting has been
updated? It would be interesting to see the output of hot threads[1] to see
what your node is doing.

[1]
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/cluster-nodes-hot-threads.html


On Fri, Apr 4, 2014 at 7:18 PM, Elliott Bradshaw ebradsh...@gmail.comwrote:

 Yes. I have run max_num_segments=1 every time.


 On Fri, Apr 4, 2014 at 12:26 PM, Michael Sick 
 michael.s...@serenesoftware.com wrote:

 Have you tried max_num_segments=1 on your optimize?

 On Fri, Apr 4, 2014 at 11:27 AM, Elliott Bradshaw 
 ebradsh...@gmail.comwrote:

 Any thoughts on this?  I've run optimize several more times, and the
 number of segments falls each time, but I'm still over 1000 segments per
 shard.  Has anyone else run into something similar?


 On Thursday, April 3, 2014 11:21:29 AM UTC-4, Elliott Bradshaw wrote:

 OK.  Optimize finally returned, so I suppose something was happening in
 the background, but I'm still seeing over 6500 segments.  Even after
 setting max_num_segments=5.  Does this seem right?  Queries are a little
 faster (350-400ms) but still not great.  Bigdesk is still showing a fair
 amount of file IO.

 On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node
 cluster, each with 64G of ram, with 24G allocated to Elasticsearch on
 each.  I've batch loaded approximately 86 million documents into a single
 index (4 shards) and have started benchmarking cross_field/multi_match
 queries on them.  The index has one replica and takes up a total of 111G.
 I've run several batches of warming queries, but queries are not as fast 
 as
 I had hoped, approximately 400-500ms each.  Given that *top *(on
 Centos) shows 5-8 GB of free memory on each server, I would assume that 
 the
 entire index has been paged into memory (I had worried about disk
 performance previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is
 composed of  7000 segments.  This seemed high to me, but maybe it's
 appropriate.  Regardless, I dispatched an optimize command, but I am not
 seeing any progress and the command has not returned.  Current merges
 remains at zero, and the segment count is not changing.  Checking out hot
 threads in ElasticHQ, I initially saw an optimize call in the stack that
 was blocked on a waitForMerge call.  This however has disappeared, and I'm
 seeing no evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else had
 similar issues.  This is the second time I have tried to optimize an index
 since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott

  --
 You received this message because you are subscribed to the Google
 Groups elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send
 an email to elasticsearch+unsubscr...@googlegroups.com.
  To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.comhttps://groups.google.com/d/msgid/elasticsearch/5391291f-5c5e-4088-a1f2-93272beef0bb%40googlegroups.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


  --
 You received this message because you are subscribed to a topic in the
 Google Groups elasticsearch group.
 To unsubscribe from this topic, visit
 https://groups.google.com/d/topic/elasticsearch/kqTRRADQBwc/unsubscribe.
 To unsubscribe from this group and all its topics, send an email to
 elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/CAP8axnD7BUziGct2%3Db%3DfupaKYFnA5fR2TBsxHoURJumHSyODFA%40mail.gmail.comhttps://groups.google.com/d/msgid/elasticsearch/CAP8axnD7BUziGct2%3Db%3DfupaKYFnA5fR2TBsxHoURJumHSyODFA%40mail.gmail.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.


  --
 You received this message because you are subscribed to the Google Groups
 elasticsearch group.
 To unsubscribe from this group and stop receiving emails from it, send an
 email to elasticsearch+unsubscr...@googlegroups.com.
 To view this discussion on the web visit
 https://groups.google.com/d/msgid/elasticsearch/CAGCt%2BFvoSQTvv%2B6G%3D3GOX27AuYdEwLiW%3Demc0JTouT9%2BBeUk_A%40mail.gmail.comhttps://groups.google.com/d/msgid/elasticsearch/CAGCt%2BFvoSQTvv%2B6G%3D3GOX27AuYdEwLiW%3Demc0JTouT9%2BBeUk_A%40mail.gmail.com?utm_medium=emailutm_source=footer
 .

 For more options, visit https://groups.google.com/d/optout.




-- 
Adrien Grand

-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to 

Elasticsearch 1.1.0 - Optimize broken?

2014-04-03 Thread Elliott Bradshaw
Hi All,

I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node cluster, 
each with 64G of ram, with 24G allocated to Elasticsearch on each.  I've 
batch loaded approximately 86 million documents into a single index (4 
shards) and have started benchmarking cross_field/multi_match queries on 
them.  The index has one replica and takes up a total of 111G.  I've run 
several batches of warming queries, but queries are not as fast as I had 
hoped, approximately 400-500ms each.  Given that *top *(on Centos) shows 
5-8 GB of free memory on each server, I would assume that the entire index 
has been paged into memory (I had worried about disk performance 
previously, as we are working in a virtualized environment).

A stats query on the index in questions shows that the index is composed of 
 7000 segments.  This seemed high to me, but maybe it's appropriate.  
Regardless, I dispatched an optimize command, but I am not seeing any 
progress and the command has not returned.  Current merges remains at zero, 
and the segment count is not changing.  Checking out hot threads in 
ElasticHQ, I initially saw an optimize call in the stack that was blocked 
on a waitForMerge call.  This however has disappeared, and I'm seeing no 
evidence that the optimize is occuring.

Does any of this seem out of the norm or unusual?  Has anyone else had 
similar issues.  This is the second time I have tried to optimize an index 
since upgrading.  I've gotten the same result both time.

Thanks in advance for any help/tips!

- Elliott

-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/aafe-0a07-4991-bcb8-8c5f94118eb7%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.


Re: Elasticsearch 1.1.0 - Optimize broken?

2014-04-03 Thread Elliott Bradshaw
OK.  Optimize finally returned, so I suppose something was happening in the 
background, but I'm still seeing over 6500 segments.  Even after setting 
max_num_segments=5.  Does this seem right?  Queries are a little faster 
(350-400ms) but still not great.  Bigdesk is still showing a fair amount of 
file IO.

On Thursday, April 3, 2014 8:47:32 AM UTC-4, Elliott Bradshaw wrote:

 Hi All,

 I've recently upgraded to Elasticsearch 1.1.0.  I've got a 4 node cluster, 
 each with 64G of ram, with 24G allocated to Elasticsearch on each.  I've 
 batch loaded approximately 86 million documents into a single index (4 
 shards) and have started benchmarking cross_field/multi_match queries on 
 them.  The index has one replica and takes up a total of 111G.  I've run 
 several batches of warming queries, but queries are not as fast as I had 
 hoped, approximately 400-500ms each.  Given that *top *(on Centos) shows 
 5-8 GB of free memory on each server, I would assume that the entire index 
 has been paged into memory (I had worried about disk performance 
 previously, as we are working in a virtualized environment).

 A stats query on the index in questions shows that the index is composed 
 of  7000 segments.  This seemed high to me, but maybe it's appropriate.  
 Regardless, I dispatched an optimize command, but I am not seeing any 
 progress and the command has not returned.  Current merges remains at zero, 
 and the segment count is not changing.  Checking out hot threads in 
 ElasticHQ, I initially saw an optimize call in the stack that was blocked 
 on a waitForMerge call.  This however has disappeared, and I'm seeing no 
 evidence that the optimize is occuring.

 Does any of this seem out of the norm or unusual?  Has anyone else had 
 similar issues.  This is the second time I have tried to optimize an index 
 since upgrading.  I've gotten the same result both time.

 Thanks in advance for any help/tips!

 - Elliott


-- 
You received this message because you are subscribed to the Google Groups 
elasticsearch group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/01ad85fc-3ad0-4c90-9f6e-1ed275bd7312%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.