Sorry for bumping this, but I'm a little stumped here.

We have some nodes that are evicting fielddata cache entries for seemingly 
no reason:
1) we've set indices.fielddata.cache.size to 10gb
2) the metrics from the node stats endpoint show that the 
indices.fielddata.memory_size_in_bytes never exceeded 3.6GB on any node.
3) the rate of eviction is normally 0, but goes up above that eventhough 
the fielddata cache size is nowhere near 10GB

Attached is a plot of the max(indices.fielddata.memory_size_in_bytes) (red 
line) and sum(indices.fielddata.evictions) (green line) across all nodes in 
the cluster. Note that we create a fresh new index every day that replaces 
an older one (that explains the change in profile around midnight).

As you can see, the size (on any given node) never exceeds 3.6GB, yet even 
at a lower value (around 2.2GB), some nodes start evicting entries from the 
cache. Also, starting around Tue 8AM, the max(field cache size) becomes 
erratic and jumps up and down.

I can't explain this behaviour, especially since we've been operating for a 
while at this volume and rate of documents. This was not happening before. 
Though it's possible that we're getting a higher volume of data, it doesn't 
look substantially different from the past.

Under what circumstances will an ES node evict entries from it's field data 
cache? We're also deleting documents from the index, can this have an 
impact? What other things should I be looking it to find a correlation (GC 
time does not seem to be correlated)?

Thanks,
Philippe

On Friday, September 12, 2014 9:33:16 AM UTC-4, Philippe Laflamme wrote:
>
> Forgot to mention that we're using ES 1.1.1
>
> On Friday, September 12, 2014 9:21:23 AM UTC-4, Philippe Laflamme wrote:
>>
>> Hi,
>>
>> I have a cluster with nodes configured with a 18G heap. We've noticed a 
>> degradation in performance recently after increasing the volume of data 
>> we're indexing.
>>
>> I think the issue is due to the field data cache doing eviction. Some 
>> nodes are doing lots of them, some aren't doing any. This is explained by 
>> our routing strategy which results in non-uniform document distribution. 
>> Maybe we can improve this eventually, but in the meantime, I'm trying to 
>> understand why the nodes are evicting cached data.
>>
>> The metrics show that the field data cache is only ~1.5GB in size, yet we 
>> have this in our elasticsearch.yml:
>>
>> indices.fielddata.cache.size: 10gb
>>
>> Why would a node evict cache entries when it should still have plenty of 
>> room to store more? Are we missing another setting? Is there a way to tell 
>> what the actual fielddata cache size is at runtime (maybe it did not pickup 
>> the configuration setting for some reason)?
>>
>> Thanks,
>> Philippe
>>
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/eafa1b0a-dbd6-4127-94d5-3733a3067bc7%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to