Re: Reduce Number of Segments
Mike, OK, thanks. I had looked at setting index.merge.scheduler.max_thread_count in the past but thought the default was already 1; I see now that it was increased. Thanks for the heads-up! I realized after sending my last reply that I’m actually getting about a 12MB/sec per server (6MB/sec per node); I failed to sum the rate for the 2 nodes on each server. With throttling disabled altogether, I did notice it jump up to about 14MB/sec, but it wasn’t substantial. Any other recommendations? I’m still seeing a significant number of segments (~ 400) created for my ‘high’ daily index. FWIW, I have another 4 servers on order which should help the situation, but I want to make sure I’m taking full advantage of my resources. Thanks, Chris From: Michael McCandless m...@elasticsearch.com Reply: elasticsearch@googlegroups.com elasticsearch@googlegroups.com Date: August 26, 2014 at 4:27:31 PM To: elasticsearch@googlegroups.com elasticsearch@googlegroups.com Subject: Re: Reduce Number of Segments OK, I would suggest setting index.merge.scheduler.max_thread_count to 1 for spinning disks. Maybe try also disabling merge throttling and see if that has an effect? 6 MB/sec seems slow... Mike McCandless http://blog.mikemccandless.com On Mon, Aug 25, 2014 at 8:57 PM, Chris Decker ch...@chris-decker.com wrote: Mike, Thanks for the response. I'm running ES 1.2.1. It appears the issue that you reported / corrected was included with ES 1.2.0. Any other ideas / suggestions? Were the settings that I posted sane? Thanks!, Chris On Monday, August 25, 2014 1:52:46 PM UTC-4, Michael McCandless wrote: Which version of ES are you using? Versions before 1.2 have a bug that caused merge throttling to throttle far more than requested such that you couldn't get any faster than ~8 MB / sec. See https://github.com/elasticsearch/elasticsearch/issues/6018 Tiered merge policy is best. Mike McCandless http://blog.mikemccandless.com On Mon, Aug 25, 2014 at 1:08 PM, Chris Decker ch...@chris-decker.com wrote: All, I’m looking for advice on how to reduce the number of segments for my indices because in my use case (log analysis), quick searches are more important than real-time access to data. I've turned many of the knobs available within ES, and read many blog postings, ES documentation, etc., but still feel like there is room for important. Specific questions I have: 1. How can I increase the current merge rate? According to Elastic HQ, my merge rate is 6 MB/s (according to Elastic HQ). I know I don't have SSDs, but with 15k drives it seems like I should be able to get better rates. I tried increasing indices.store.throttle.max_bytes_per_sec from the default of 20mb to 40mb in my templates, but I didn't see a noticeable change in disk IOps or the merge rate the next day. Did I do something incorrectly? I'm going to experiment with setting it overall with index.store.throttle.max_bytes_per_sec and removing it from my templates. 2. Should I move away from the default merge policy, or stick with the default (tiered)? Any advice you have is much appreciated; additional details on my situation are below. - I generate 2 indices per day - “high” and “low”. I usually end up with ~ 450 segments for my ‘high’ index (see attached), and another ~ 200 segments for my ‘low’ index, which I then optimize once I roll-over to the next day’s indices. - 4 ES servers (soon to be 8). — Each server has: 12 Xeon cores running at 2.3 GHz 15k drives 128 GB of RAM 68 GB used for OS / file system machine 60 GB used by 2 JVMs - Index ~ 750 GB per day; 1.5 TB if you include the replicas - Relevant configs: TEMPLATE: index.refresh_interval : 60s, index.number_of_replicas : 1, index.number_of_shards : 4, index.merge.policy.max_merged_segment : 50g, index.merge.policy.segments_per_tier : 5, index.merge.policy.max_merge_at_once : “5”, indices.store.throttle.max_bytes_per_sec : 40mb. ELASTICSEARCH.YML: indices.memory.index_buffer_size: 30% Thanks in advance!, Chris -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearc...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/002cb4cc-fa2e-43c3-b2d3-29580742c91a%40googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/46ecc658-502f-46c7-b2b9-db9fd0e9f58f%40googlegroups.com. For more options, visit https://groups.google.com/d/optout. -- You
Re: Reduce Number of Segments
Mike, :) I upgraded to 1.3.2 yesterday mid-afternoon. So far things feel much snappier, but I wiped my ‘data’ directory so ES has less to search (though most of my queries only go back 1 day anyways; I go back 3 days on Monday’s to account for the weekend days I missed though). I wish I could get SSDs. The good news is that when I get these 4 additional servers, ES will have a total of 248 GB of RAM, which should allow quite a bit of data to be cached. Do you have any guidelines for using warmers? From my understanding, they help “warm up” the segments, so I have them defined in my template so that the common term queries are executed automatically. I expected this to make it so that when I hit a page that used those exact term queries, the data would pretty much load instantly because it was in the cache. It doesn’t appear they load instantly, though the warmers have seemed to help. Thanks, Chris From: Michael McCandless m...@elasticsearch.com Reply: Michael McCandless m...@elasticsearch.com Date: August 28, 2014 at 3:21:30 PM To: Chris Decker ch...@chris-decker.com Cc: elasticsearch@googlegroups.com elasticsearch@googlegroups.com Subject: Re: Reduce Number of Segments Only things I can think of are to upgrade to 1.3.2, and switch to SSDs :) Mike McCandless http://blog.mikemccandless.com On Thu, Aug 28, 2014 at 12:23 PM, Chris Decker ch...@chris-decker.com wrote: Mike, OK, thanks. I had looked at setting index.merge.scheduler.max_thread_count in the past but thought the default was already 1; I see now that it was increased. Thanks for the heads-up! I realized after sending my last reply that I’m actually getting about a 12MB/sec per server (6MB/sec per node); I failed to sum the rate for the 2 nodes on each server. With throttling disabled altogether, I did notice it jump up to about 14MB/sec, but it wasn’t substantial. Any other recommendations? I’m still seeing a significant number of segments (~ 400) created for my ‘high’ daily index. FWIW, I have another 4 servers on order which should help the situation, but I want to make sure I’m taking full advantage of my resources. Thanks, Chris From: Michael McCandless m...@elasticsearch.com Reply: elasticsearch@googlegroups.com elasticsearch@googlegroups.com Date: August 26, 2014 at 4:27:31 PM To: elasticsearch@googlegroups.com elasticsearch@googlegroups.com Subject: Re: Reduce Number of Segments OK, I would suggest setting index.merge.scheduler.max_thread_count to 1 for spinning disks. Maybe try also disabling merge throttling and see if that has an effect? 6 MB/sec seems slow... Mike McCandless http://blog.mikemccandless.com On Mon, Aug 25, 2014 at 8:57 PM, Chris Decker ch...@chris-decker.com wrote: Mike, Thanks for the response. I'm running ES 1.2.1. It appears the issue that you reported / corrected was included with ES 1.2.0. Any other ideas / suggestions? Were the settings that I posted sane? Thanks!, Chris On Monday, August 25, 2014 1:52:46 PM UTC-4, Michael McCandless wrote: Which version of ES are you using? Versions before 1.2 have a bug that caused merge throttling to throttle far more than requested such that you couldn't get any faster than ~8 MB / sec. See https://github.com/elasticsearch/elasticsearch/issues/6018 Tiered merge policy is best. Mike McCandless http://blog.mikemccandless.com On Mon, Aug 25, 2014 at 1:08 PM, Chris Decker ch...@chris-decker.com wrote: All, I’m looking for advice on how to reduce the number of segments for my indices because in my use case (log analysis), quick searches are more important than real-time access to data. I've turned many of the knobs available within ES, and read many blog postings, ES documentation, etc., but still feel like there is room for important. Specific questions I have: 1. How can I increase the current merge rate? According to Elastic HQ, my merge rate is 6 MB/s (according to Elastic HQ). I know I don't have SSDs, but with 15k drives it seems like I should be able to get better rates. I tried increasing indices.store.throttle.max_bytes_per_sec from the default of 20mb to 40mb in my templates, but I didn't see a noticeable change in disk IOps or the merge rate the next day. Did I do something incorrectly? I'm going to experiment with setting it overall with index.store.throttle.max_bytes_per_sec and removing it from my templates. 2. Should I move away from the default merge policy, or stick with the default (tiered)? Any advice you have is much appreciated; additional details on my situation are below. - I generate 2 indices per day - “high” and “low”. I usually end up with ~ 450 segments for my ‘high’ index (see attached), and another ~ 200 segments for my ‘low’ index, which I then optimize once I roll-over to the next day’s indices. - 4 ES servers (soon to be 8). — Each server has: 12 Xeon cores running at 2.3 GHz 15k drives 128 GB of RAM 68 GB
Reduce Number of Segments
All, I’m looking for advice on how to reduce the number of segments for my indices because in my use case (log analysis), quick searches are more important than real-time access to data. I've turned many of the knobs available within ES, and read many blog postings, ES documentation, etc., but still feel like there is room for important. Specific questions I have: 1. How can I increase the current merge rate? According to Elastic HQ, my merge rate is 6 MB/s (according to Elastic HQ). I know I don't have SSDs, but with 15k drives it seems like I should be able to get better rates. I tried increasing indices.store.throttle.max_bytes_per_sec from the default of 20mb to 40mb in my templates, but I didn't see a noticeable change in disk IOps or the merge rate the next day. Did I do something incorrectly? I'm going to experiment with setting it overall with index.store.throttle.max_bytes_per_sec and removing it from my templates. 2. Should I move away from the default merge policy, or stick with the default (tiered)? Any advice you have is much appreciated; additional details on my situation are below. - I generate 2 indices per day - “high” and “low”. I usually end up with ~ 450 segments for my ‘high’ index (see attached), and another ~ 200 segments for my ‘low’ index, which I then optimize once I roll-over to the next day’s indices. - 4 ES servers (soon to be 8). — Each server has: 12 Xeon cores running at 2.3 GHz 15k drives 128 GB of RAM 68 GB used for OS / file system machine 60 GB used by 2 JVMs - Index ~ 750 GB per day; 1.5 TB if you include the replicas - Relevant configs: TEMPLATE: index.refresh_interval : 60s, index.number_of_replicas : 1, index.number_of_shards : 4, index.merge.policy.max_merged_segment : 50g, index.merge.policy.segments_per_tier : 5, index.merge.policy.max_merge_at_once : “5”, indices.store.throttle.max_bytes_per_sec : 40mb. ELASTICSEARCH.YML: indices.memory.index_buffer_size: 30% Thanks in advance!, Chris -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/002cb4cc-fa2e-43c3-b2d3-29580742c91a%40googlegroups.com. For more options, visit https://groups.google.com/d/optout. index shard prirep ipsegment generation docs.count docs.deleted size size.memory committed searchable version compound high-2014.08.24 0 p 1.2.3.4 _lrk 2820858151130 3.9gb55697309 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _10g547237 190555190 12.6gb 178323656 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _1l2673950 166506980 9.6gb 155076791 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _1suo8404842598150 2.4gb40337901 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _1xt39047147413940 2.9gb44799302 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _24y59972582418050 5.3gb79225058 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _27hs 10302434323200 2.3gb33132147 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2ccn 10931952103010 3.6gb49775842 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2frr 113751 101417040 6.5gb97469720 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2iks 11738821203220 1.5gb20871023 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2jvj 11907126549990 1.9gb26059406 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2nc6 12355859417140 4.2gb56767385 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2oid 12507711019720 833.2mb11028090 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2pbu 12613814358020 1gb14298968 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2qlr 12779123866600 1.7gb23440373 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2r4e 128462 6356460 485.2mb 7910210 true true 4.8 false high-2014.08.24 0 p 1.2.3.4 _2rlz 12909511333970
Re: Reduce Number of Segments
Mike, Thanks for the response. I'm running ES 1.2.1. It appears the issue that you reported / corrected was included with ES 1.2.0. *Any other ideas / suggestions? *Were the settings that I posted sane? Thanks!, Chris On Monday, August 25, 2014 1:52:46 PM UTC-4, Michael McCandless wrote: Which version of ES are you using? Versions before 1.2 have a bug that caused merge throttling to throttle far more than requested such that you couldn't get any faster than ~8 MB / sec. See https://github.com/elasticsearch/elasticsearch/issues/6018 Tiered merge policy is best. Mike McCandless http://blog.mikemccandless.com On Mon, Aug 25, 2014 at 1:08 PM, Chris Decker ch...@chris-decker.com javascript: wrote: All, I’m looking for advice on how to reduce the number of segments for my indices because in my use case (log analysis), quick searches are more important than real-time access to data. I've turned many of the knobs available within ES, and read many blog postings, ES documentation, etc., but still feel like there is room for important. Specific questions I have: 1. How can I increase the current merge rate? According to Elastic HQ, my merge rate is 6 MB/s (according to Elastic HQ). I know I don't have SSDs, but with 15k drives it seems like I should be able to get better rates. I tried increasing indices.store.throttle.max_bytes_per_sec from the default of 20mb to 40mb in my templates, but I didn't see a noticeable change in disk IOps or the merge rate the next day. Did I do something incorrectly? I'm going to experiment with setting it overall with index.store.throttle.max_bytes_per_sec and removing it from my templates. 2. Should I move away from the default merge policy, or stick with the default (tiered)? Any advice you have is much appreciated; additional details on my situation are below. - I generate 2 indices per day - “high” and “low”. I usually end up with ~ 450 segments for my ‘high’ index (see attached), and another ~ 200 segments for my ‘low’ index, which I then optimize once I roll-over to the next day’s indices. - 4 ES servers (soon to be 8). — Each server has: 12 Xeon cores running at 2.3 GHz 15k drives 128 GB of RAM 68 GB used for OS / file system machine 60 GB used by 2 JVMs - Index ~ 750 GB per day; 1.5 TB if you include the replicas - Relevant configs: TEMPLATE: index.refresh_interval : 60s, index.number_of_replicas : 1, index.number_of_shards : 4, index.merge.policy.max_merged_segment : 50g, index.merge.policy.segments_per_tier : 5, index.merge.policy.max_merge_at_once : “5”, indices.store.throttle.max_bytes_per_sec : 40mb. ELASTICSEARCH.YML: indices.memory.index_buffer_size: 30% Thanks in advance!, Chris -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearc...@googlegroups.com javascript:. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/002cb4cc-fa2e-43c3-b2d3-29580742c91a%40googlegroups.com https://groups.google.com/d/msgid/elasticsearch/002cb4cc-fa2e-43c3-b2d3-29580742c91a%40googlegroups.com?utm_medium=emailutm_source=footer . For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/46ecc658-502f-46c7-b2b9-db9fd0e9f58f%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.
Re: Kibana Filtered Queries Failing After ES 1.0 Upgrade
Yes - I downloaded the master from GitHub and was running still seeing the issue. On Friday, February 14, 2014 10:38:49 AM UTC-5, Binh Ly wrote: I'm curious if you're running the latest Kibana, or an older one. The field query has been deprecated (and removed in ES 1.0) which is the cause of your error: http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/_deprecations.html -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/a4a88f67-e5b8-4945-ad56-586ac5b34b48%40googlegroups.com. For more options, visit https://groups.google.com/groups/opt_out.
Re: Kibana Filtered Queries Failing After ES 1.0 Upgrade
Note that I'm using the past tense only because I reverted back to ES 0.90.9, not because I figured out how to solve the issue :) -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/7c9221a1-0866-4e46-a345-92d99cc176d0%40googlegroups.com. For more options, visit https://groups.google.com/groups/opt_out.
Re: Kibana Filtered Queries Failing After ES 1.0 Upgrade
Binh, First let me thank you for helping me track down what’s going on here. So I can confirm that everything I see is the same as what you saw (with the exception that at the end of the Kibana version mine says [master]). If I enter a query for: querystring must query : action:connect …then the expected results come back. Try this now: Go to your logstash dashboard and click one of the rows. Click the magnifying glass near one of the fields and see if you get back results. The above is how my users are the primary way that my users add filters to find what they’re looking for, and that’s the part that isn’t working for me. Thanks, Chris -- You received this message because you are subscribed to the Google Groups elasticsearch group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/etPan.52fe5cf3.79e2a9e3.e8%40april.sos.its.psu.edu. For more options, visit https://groups.google.com/groups/opt_out.