Hi,, The only things I see is ES_HEAP_SIZE with 50% this can bet set to 8gb. I'm seeing some swap usage, you could disable swap completely.
hth,, A. On Monday, January 5, 2015 12:48:53 PM UTC+1, Darren Mansell wrote: > > Hi all, > > We have a 4 node VM dev cluster with 3 data nodes and 1 master node. The 3 > data nodes are taking lots of CPU and the load average is high when the > servers are sitting idle with no accesses or indexing. > > This seemed to be fine before we updated to 1.4.2 before the new year, but > I can't confirm this is definitely the reason. > > The hot threads from each node seem to indicate something to do with the > filesystem most of the time e.g. > > 87.7% (438.6ms out of 500ms) cpu usage by thread > 'elasticsearch[potassium][management][T#4]' > 2/10 snapshots sharing following 21 elements > org.apache.lucene.store.FSDirectory.listAll(FSDirectory.java:223) > org.apache.lucene.store.FSDirectory.listAll(FSDirectory.java:242) > > org.apache.lucene.store.FileSwitchDirectory.listAll(FileSwitchDirectory.java:87) > > org.apache.lucene.store.FilterDirectory.listAll(FilterDirectory.java:48) > > org.elasticsearch.index.store.DistributorDirectory.listAll(DistributorDirectory.java:88) > > org.apache.lucene.store.FilterDirectory.listAll(FilterDirectory.java:48) > > org.elasticsearch.common.lucene.Directories.estimateSize(Directories.java:40) > org.elasticsearch.index.store.Store.stats(Store.java:216) > > org.elasticsearch.index.shard.service.InternalIndexShard.storeStats(InternalIndexShard.java:540) > > org.elasticsearch.action.admin.indices.stats.CommonStats.<init>(CommonStats.java:134) > > org.elasticsearch.action.admin.indices.stats.ShardStats.<init>(ShardStats.java:49) > > org.elasticsearch.indices.InternalIndicesService.stats(InternalIndicesService.java:212) > org.elasticsearch.node.service.NodeService.stats(NodeService.java:156) > > org.elasticsearch.action.admin.cluster.node.stats.TransportNodesStatsAction.nodeOperation(TransportNodesStatsAction.java:96) > > org.elasticsearch.action.admin.cluster.node.stats.TransportNodesStatsAction.nodeOperation(TransportNodesStatsAction.java:44) > > org.elasticsearch.action.support.nodes.TransportNodesOperationAction$NodeTransportHandler.messageReceived(TransportNodesOperationAction.java:278) > > org.elasticsearch.action.support.nodes.TransportNodesOperationAction$NodeTransportHandler.messageReceived(TransportNodesOperationAction.java:269) > > org.elasticsearch.transport.netty.MessageChannelHandler$RequestHandler.run(MessageChannelHandler.java:275) > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > java.lang.Thread.run(Thread.java:745) > > > Please also see the image attachment for the high CPU and load. > > > The VMs are on ESXi 5.5 on 2 * 4 core HT Xeon E5530 so server hardware is > pretty heavy. The nodes are set up with 16GB RAM, 2 vCPUs, all ES defaults > apart from: > > > [root@potassium ~]# grep -v ^# /etc/elasticsearch/elasticsearch.yml | grep -v > ^$cluster.name: es-devnode.name: "potassium" > node.master: false > node.data: true > path.data: /data/elasticsearch > marvel.agent.exporter.es.hosts: ["hydrogen:9200"] > > > and > > > [root@potassium ~]# grep -v ^# /etc/sysconfig/elasticsearch | grep -v ^$ > ES_HOME=/usr/share/elasticsearch > ES_HEAP_SIZE=6g > MAX_OPEN_FILES=65535 > MAX_MAP_COUNT=262144 > LOG_DIR=/var/log/elasticsearch > DATA_DIR=/var/lib/elasticsearch > WORK_DIR=/tmp/elasticsearch > CONF_DIR=/etc/elasticsearch > CONF_FILE=/etc/elasticsearch/elasticsearch.yml > ES_USER=elasticsearch > > > So far I've tried: > > > > - Dropping all data and loading again using logstash > - Deleting XFS filesystem and changing to ext4 > - Removing all plugins > - Leaving for about 2 weeks in case it was doing background optimisation > - and various other things > > Does anyone have any suggestions about where I should look next, or any > thoughts about what could be happening? Please let me know if I can pull any > other info off the nodes or cluster to help diagnose. > > Many thanks, > Darren > > -- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/9447a9cc-f65b-4b46-8a0c-35002368c889%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.