Workers do not restart on their own.  There *must* be supervisor logs about
starting workers and noticing workers are dead.  Look through the
supervisor logs, don't just search for errors.

On Thu, Dec 15, 2016 at 2:13 AM Mostafa Gomaa <[email protected]> wrote:

> That STDERR is most probably a dependency issue. You are probably using
> multiple external libraries that include conflicting logging libraries as
> dependencies. As for the changing PID, I am not too sure about that, but I
> am interested to see if someone else can shed some light on that.
>
> On Thu, Dec 15, 2016 at 11:57 AM, Eranga Heshan <[email protected]>
> wrote:
>
> I am using storm-1.0.2.
>
> I checked all the logs in STORM_DIR\logs. I found no [ERROR] logs printed
> only [INFO] logs. I saw STDERR logs as [INFO]. For an example,
>
> STDERR [INFO] SLF4J: Class path contains multiple SLF4J bindings.
>
> There was no clear log written to identify that the worker was killed. As
> I mentioned before, I am not sure actually what happened to the worker. All
> I observed was that the worker changed its PID while the topology was still
> running. But ultimately, my topology ran fine and also produced the desired
> output.
>
> Are you sure that there is an issue to be fixed?
>
> Thanks,
> Regards,
>
>
>
> Eranga Heshan
> *Undergraduate*
> Computer Science & Engineering
> University of Moratuwa
> Mobile:  +94 71 138 2686 <%2B94%2071%20552%202087>
> Email: [email protected] <[email protected]>
> <https://www.facebook.com/erangaheshan>
> <https://twitter.com/erangaheshan>
> <https://www.linkedin.com/in/erangaheshan>
>
>
>
> On Thu, Dec 15, 2016 at 8:27 AM, Erik Weathers <[email protected]>
> wrote:
>
> What version of storm are you running?  The newer ones (I believe 0.10+)
> record stdout/stderr from workers via a wrapping "LogWriter" process.  If
> your worker process is dying as you say it is, there *will* be logs in 1 or
> more of these places:
>
>    - supervisor logs
>    - worker logs
>    - worker stderr/stdout logs
>
> You should figure out why it's dying and fix whatever that issue is.
>
> - Erik
>
> On Wed, Dec 14, 2016 at 6:44 PM, Eranga Heshan <[email protected]>
> wrote:
>
> I checked the log files and there are no errors logged.
>
> While running the topology I checked that log directory. Although the
> worker gets restarted, it writes the log to the same file as long as the
> new worker runs on the same port (port 6704). In my case, after a while, it
> selects another port (port 6700). Then it writes a new log. (Log directory
> is named after the port number)
>
> I would like to know if this is a normal behavior of storm worker. Because
> this scenario does not affect the topology process.
>
> Thanks,
> Regards,
>
>
> Eranga Heshan
> *Undergraduate*
> Computer Science & Engineering
> University of Moratuwa
> Mobile:  +94 71 138 2686 <%2B94%2071%20552%202087>
> Email: [email protected] <[email protected]>
> <https://www.facebook.com/erangaheshan>
> <https://twitter.com/erangaheshan>
> <https://www.linkedin.com/in/erangaheshan>
>
>
>
> On Wed, Dec 14, 2016 at 7:08 PM, Mostafa Gomaa <[email protected]> wrote:
>
> I would check the log file for that worker. You can find it
> storm/logs/workers-artifacts/topology_id. check if there are any errors
> that are causing the worker to restart.
>
> On Wed, Dec 14, 2016 at 3:36 PM, Eranga Heshan <[email protected]>
> wrote:
>
> Hi all,
>
> I recently ran a topology which consists of 5 workers on 4 node cluster.
> Every worker has the below configuration parameter set.
>
> worker.childopts: "-Xms1500m"
>
> When the topology is submitted, I checked for each worker's behavior and
> found out that one worker (runs alone in one node) keeps restarting.
>
> It actually doesn't affect the process because the restarted worker does
> the same job like the previous. But I am curious to know what exactly is
> happening to the worker to get restarted.
>
> I checked the free memory of that particular worker's node continuously
> and found out that it gets restarted even it has enough memory left (more
> than 1GB). However, there might be many events buffered to be processed by
> that worker since the spout is producing events at a much higher rate.
>
> Given the above details can anyone please give me a clarification on what
> would be happening to the worker?
>
> Thanks,
> Regards,
>
>
>
>
>
> Eranga Heshan
> *Undergraduate*
> Computer Science & Engineering
> University of Moratuwa
> Mobile:  +94 71 138 2686 <%2B94%2071%20552%202087>
> Email: [email protected] <[email protected]>
> <https://www.facebook.com/erangaheshan>
> <https://twitter.com/erangaheshan>
> <https://www.linkedin.com/in/erangaheshan>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>

Reply via email to