I am using storm-1.0.2.

I checked all the logs in STORM_DIR\logs. I found no [ERROR] logs printed
only [INFO] logs. I saw STDERR logs as [INFO]. For an example,

STDERR [INFO] SLF4J: Class path contains multiple SLF4J bindings.

There was no clear log written to identify that the worker was killed. As I
mentioned before, I am not sure actually what happened to the worker. All I
observed was that the worker changed its PID while the topology was still
running. But ultimately, my topology ran fine and also produced the desired
output.

Are you sure that there is an issue to be fixed?

Thanks,
Regards,



Eranga Heshan
*Undergraduate*
Computer Science & Engineering
University of Moratuwa
Mobile:  +94 71 138 2686 <%2B94%2071%20552%202087>
Email: [email protected] <[email protected]>
<https://www.facebook.com/erangaheshan>   <https://twitter.com/erangaheshan>
   <https://www.linkedin.com/in/erangaheshan>

On Thu, Dec 15, 2016 at 8:27 AM, Erik Weathers <[email protected]>
wrote:

> What version of storm are you running?  The newer ones (I believe 0.10+)
> record stdout/stderr from workers via a wrapping "LogWriter" process.  If
> your worker process is dying as you say it is, there *will* be logs in 1 or
> more of these places:
>
>    - supervisor logs
>    - worker logs
>    - worker stderr/stdout logs
>
> You should figure out why it's dying and fix whatever that issue is.
>
> - Erik
>
> On Wed, Dec 14, 2016 at 6:44 PM, Eranga Heshan <[email protected]>
> wrote:
>
>> I checked the log files and there are no errors logged.
>>
>> While running the topology I checked that log directory. Although the
>> worker gets restarted, it writes the log to the same file as long as the
>> new worker runs on the same port (port 6704). In my case, after a while, it
>> selects another port (port 6700). Then it writes a new log. (Log directory
>> is named after the port number)
>>
>> I would like to know if this is a normal behavior of storm worker.
>> Because this scenario does not affect the topology process.
>>
>> Thanks,
>> Regards,
>>
>>
>> Eranga Heshan
>> *Undergraduate*
>> Computer Science & Engineering
>> University of Moratuwa
>> Mobile:  +94 71 138 2686 <%2B94%2071%20552%202087>
>> Email: [email protected] <[email protected]>
>> <https://www.facebook.com/erangaheshan>
>> <https://twitter.com/erangaheshan>
>> <https://www.linkedin.com/in/erangaheshan>
>>
>> On Wed, Dec 14, 2016 at 7:08 PM, Mostafa Gomaa <[email protected]>
>> wrote:
>>
>>> I would check the log file for that worker. You can find it
>>> storm/logs/workers-artifacts/topology_id. check if there are any errors
>>> that are causing the worker to restart.
>>>
>>> On Wed, Dec 14, 2016 at 3:36 PM, Eranga Heshan <[email protected]>
>>> wrote:
>>>
>>>> Hi all,
>>>>
>>>> I recently ran a topology which consists of 5 workers on 4 node
>>>> cluster. Every worker has the below configuration parameter set.
>>>>
>>>> worker.childopts: "-Xms1500m"
>>>>
>>>> When the topology is submitted, I checked for each worker's behavior
>>>> and found out that one worker (runs alone in one node) keeps restarting.
>>>>
>>>> It actually doesn't affect the process because the restarted worker
>>>> does the same job like the previous. But I am curious to know what exactly
>>>> is happening to the worker to get restarted.
>>>>
>>>> I checked the free memory of that particular worker's node continuously
>>>> and found out that it gets restarted even it has enough memory left (more
>>>> than 1GB). However, there might be many events buffered to be processed by
>>>> that worker since the spout is producing events at a much higher rate.
>>>>
>>>> Given the above details can anyone please give me a clarification on
>>>> what would be happening to the worker?
>>>>
>>>> Thanks,
>>>> Regards,
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Eranga Heshan
>>>> *Undergraduate*
>>>> Computer Science & Engineering
>>>> University of Moratuwa
>>>> Mobile:  +94 71 138 2686 <%2B94%2071%20552%202087>
>>>> Email: [email protected] <[email protected]>
>>>> <https://www.facebook.com/erangaheshan>
>>>> <https://twitter.com/erangaheshan>
>>>> <https://www.linkedin.com/in/erangaheshan>
>>>>
>>>
>>>
>>
>

Reply via email to