Hi Ethan Li,
Sorry for the late reply. Please find the output where it is showing -Xmx2048m for worker heap . But in storm ui we are seeing the worker allocated memory as 65MB for each worker. java -server -Dlogging.sensitivity=S3 -Dlogfile.name=worker.log -Dstorm.home=/opt/storm/apache-storm-1.2.1 -Dworkers.artifacts=/var/log/storm/workers-artifacts -Dstorm.id=Topology_334348-43-1580365369 -Dworker.id=f1e3e060-0b32-4ecd-8c34-c486258264a4 -Dworker.port=6707 -Dstorm.log.dir=/var/log/storm -Dlog4j.configurationFile=/opt/storm/apache-storm-1.2.1/log4j2/worker.xml -DLog4jContextSelector=org.apache.logging.log4j.core.selector.BasicContextSelector -Dstorm.local.dir=/var/log/storm/tmp -Xmx2048m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump -Djava.library.path=/var/log/storm/tmp/supervisor/stormdist/Topology_334348-43-1580365369/resources/Linux-amd64:/var/log/storm/tmp/supervisor/stormdist/Topology_334348-43-1580365369/resources:/usr/local/lib:/opt/local/lib:/usr/lib -Dstorm.conf.file= -Dstorm.options= -Djava.io.tmpdir=/var/log/storm/tmp/workers/f1e3e060-0b32-4ecd-8c34-c486258264a4/tmp -cp /opt/storm/apache-storm-1.2.1/lib/*:/opt/storm/apache-storm-1.2.1/extlib/*:/opt/storm/apache-storm-1.2.1/conf:/var/log/storm/tmp/supervisor/stormdist/Topology_334348-43-1580365369/stormjar.jar org.apache.storm.daemon.worker Topology_334348-43-1580365369 7fe05c2b-ebcf-491b-a8cc-2565834b5988 6707 f1e3e060-0b32-4ecd-8c34-c486258264a4 On Tue, 4 Feb 2020 at 04:36, Ethan Li <ethanopensou...@gmail.com> wrote: > > This is where the worker launch command is composed: > > https://github.com/apache/storm/blob/1.x-branch/storm-core/src/jvm/org/apache/storm/daemon/supervisor/BasicContainer.java#L653-L671 > > Since your worker.childopts is set, and topology.worker.childopts is empty, > > > worker.childopts: "-Xmx2048m -XX:+PrintGCDetails > -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps > -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 > -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError > -XX:HeapDumpPath=artifacts/heapdump” > > > The command to launch the worker process should have -Xmx2048m. > > I don’t see why it would be 65MB. And what do you mean by coming as 65MB > only? Is it only committed 65MB? Or is the max only 65MB? > > Could you submit the topology and show the result of “ps -aux |grep > --ignore-case worker”? This will show you the JVM parameters of the worker > process. > > > (BTW, -Xmx%HEAP-MEM%m in worker.childopts is designed to be replaced > https://github.com/apache/storm/blob/1.x-branch/storm-core/src/jvm/org/apache/storm/daemon/supervisor/BasicContainer.java#L392) > > > > > On Jan 30, 2020, at 2:12 AM, Narasimhan Chengalvarayan > <narasimha...@gmail.com> wrote: > > Hi Ethan, > > > Please find the configuration detail > > ********************************************** > > #Licensed to the Apache Software Foundation (ASF) under one > # or more contributor license agreements. See the NOTICE file > # distributed with this work for additional information > # regarding copyright ownership. The ASF licenses this file > # to you under the Apache License, Version 2.0 (the > # "License"); you may not use this file except in compliance > # with the License. You may obtain a copy of the License at > # > # http://www.apache.org/licenses/LICENSE-2.0 > # > # Unless required by applicable law or agreed to in writing, software > # distributed under the License is distributed on an "AS IS" BASIS, > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. > # See the License for the specific language governing permissions and > # limitations under the License. > > ########### These MUST be filled in for a storm configuration > storm.zookeeper.servers: > - "ZK1" > - "ZK2" > - "ZK3" > > nimbus.seeds: ["host1","host2"] > ui.port : 8081 > storm.log.dir: "/var/log/storm" > storm.local.dir: "/var/log/storm/tmp" > supervisor.slots.ports: > - 6700 > - 6701 > - 6702 > - 6703 > - 6704 > - 6705 > - 6706 > - 6707 > - 6708 > - 6709 > - 6710 > - 6711 > - 6712 > - 6713 > - 6714 > - 6715 > - 6716 > - 6717 > worker.heap.memory.mb: 1639 > topology.worker.max.heap.size.mb: 1639 > worker.childopts: "-Xmx2048m -XX:+PrintGCDetails > -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps > -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 > -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError > -XX:HeapDumpPath=artifacts/heapdump" > worker.gc.childopts: "" > > topology.min.replication.count: 2 > # > # > # ##### These may optionally be filled in: > # > ## List of custom serializations > # topology.kryo.register: > # - org.mycompany.MyType > # - org.mycompany.MyType2: org.mycompany.MyType2Serializer > # > ## List of custom kryo decorators > # topology.kryo.decorators: > # - org.mycompany.MyDecorator > # > ## Locations of the drpc servers > # drpc.servers: > # - "server1" > # - "server2" > > ## Metrics Consumers > # topology.metrics.consumer.register: > # - class: "org.apache.storm.metric.LoggingMetricsConsumer" > # parallelism.hint: 1 > # - class: "org.mycompany.MyMetricsConsumer" > # parallelism.hint: 1 > # argument: > # - endpoint: "metrics-collector.mycompany.org" > > ********************************************************************************************* > > > On Thu, 30 Jan 2020 at 03:07, Ethan Li <ethanopensou...@gmail.com> wrote: > > > I am not sure. Can you provide your configs? > > > > On Jan 28, 2020, at 6:33 PM, Narasimhan Chengalvarayan > <narasimha...@gmail.com> wrote: > > Hi Team, > > Do you have any idea, In storm apache 1.1.0 we have set worker size as > 2 GB , Once we upgrade to 1.2.1 .It was coming as 65MB only. please > help us .DO we need to follow different configuration setting for > storm 1.2.1 or it is a bug. > > On Mon, 27 Jan 2020 at 16:44, Narasimhan Chengalvarayan > <narasimha...@gmail.com> wrote: > > > Hi Team, > > In storm 1.2.1 version, worker memory is showing as 65MB. But we have > set worker memory has 2GB. > > On Fri, 24 Jan 2020 at 01:25, Ethan Li <ethanopensou...@gmail.com> wrote: > > > > 1) What is stored in Workerbeats znode? > > > Worker periodically sends heartbeat to zookeeper under workerbeats node. > > 2) Which settings control the frequency of workerbeats update > > > > https://github.com/apache/storm/blob/1.x-branch/storm-core/src/jvm/org/apache/storm/Config.java#L1534-L1539 > task.heartbeat.frequency.secs Default to 3 > > 3)What will be the impact if the frequency is reduced > > > Nimbus get the worker status from workerbeat znode to know if executors on > workers are alive or not. > https://github.com/apache/storm/blob/1.x-branch/storm-core/src/jvm/org/apache/storm/Config.java#L595-L601 > If heartbeat exceeds nimbus.task.timeout.secs (default to 30), nimbus will > think the certain executor is dead and try to reschedule. > > To reduce the issue on zookeeper, a pacemaker component was introduced. > https://github.com/apache/storm/blob/master/docs/Pacemaker.md > You might want to use it too. > > Thanks > > > On Dec 10, 2019, at 4:36 PM, Surajeet Dev <surajeetapa...@gmail.com> wrote: > > We upgraded Storm version to 1.2.1 , and since then have been consistently > observing Zookeeper session timeouts . > > On analysis , we observed that there is high frequency of updates on > workerbeats znode with data upto size of 50KB. This causes the Garbage > Collector to kickoff lasting more than 15 secs , resulting in Zookeper > session timeout > > I understand , increasing the session timeout will alleviate the issue , but > we have already done that twice > > My questions are: > > 1) What is stored in Workerbeats znode? > 2) Which settings control the frequency of workerbeats update > 3)What will be the impact if the frequency is reduced > > > > > > -- > Thanks > C.Narasimhan > 09739123245 > > > > > -- > Thanks > C.Narasimhan > 09739123245 > > > > > -- > Thanks > C.Narasimhan > 09739123245 > > -- Thanks C.Narasimhan 09739123245