Re: Yarn -- one of the daemons getting killed
Hi Vinod, Thanks for the link, I went through it and it looks like the OOM killer picks a process that has the highest oom_score. I have tried to capture oom_score for all the YARN daemon processes after each run of my application.The first time I have captured these details, I see that the name node is killed where as the Node Manager has the highest score. So, I don't if it is really the OOM killer that has killed it! Please see the output of my run attached, which also has the output of free command after each run. The output of free command doesn't either show any exhaustion of system memory. Also, one more thing I have done today is, I have added audit rules for each of the daemons to capture all the system calls. And, in the audit log, I see futex() system call occurring in the killed daemon processes. I don't know if it causes the daemon to die? and why does that call happen... Thanks, Kishore On Wed, Dec 18, 2013 at 12:31 AM, Vinod Kumar Vavilapalli < vino...@hortonworks.com> wrote: > That's good info. It is more than likely that it is the OOM killer. See > http://stackoverflow.com/questions/726690/who-killed-my-process-and-whyfor > example. > > Thanks, > +Vinod > > On Dec 17, 2013, at 1:26 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > Hi Jeff, > > I have run the resource manager in the foreground without nohup and here > are the messages when it was killed, it says it is "Killed" but doesn't say > why! > > 13/12/17 03:14:54 INFO capacity.CapacityScheduler: Application > appattempt_1387266015651_0258_01 released container > container_1387266015651_0258_01_03 on node: host: isredeng:36576 > #containers=2 available=7936 used=256 with event: FINISHED > 13/12/17 03:14:54 INFO rmcontainer.RMContainerImpl: > container_1387266015651_0258_01_05 Container Transitioned from ACQUIRED > to RUNNING > Killed > > > Thanks, > Kishore > > > On Mon, Dec 16, 2013 at 11:10 PM, Jeff Stuckman wrote: > >> What if you open the daemons in a "screen" session rather than running >> them in the background -- for example, run "yarn resourcemanager". Then you >> can see exactly when they terminate, and hopefully why. >> >> *From: *Krishna Kishore Bonagiri >> *Sent: *Monday, December 16, 2013 6:20 AM >> *To: *user@hadoop.apache.org >> *Reply To: *user@hadoop.apache.org >> *Subject: *Re: Yarn -- one of the daemons getting killed >> >> Hi Vinod, >> >> Yes, I am running on Linux. >> >> I was actually searching for a corresponding message in >> /var/log/messages to confirm that OOM killed my daemons, but could not find >> any corresponding messages there! According to the following link, it looks >> like if it is a memory issue, I should see a messages even if OOM is >> disabled, but I don't see it. >> >> http://www.redhat.com/archives/taroon-list/2007-August/msg6.html >> >>And, is memory consumption more in case of two node cluster than a >> single node one? Also, I see this problem only when I give "*" as the node >> name. >> >>One other thing I suspected was the allowed number of user processes, >> I increased that to 31000 from 1024 but that also didn't help. >> >> Thanks, >> Kishore >> >> >> On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli < >> vino...@hortonworks.com> wrote: >> >>> Yes, that is what I suspect. That is why I asked if everything is on a >>> single node. If you are running linux, linux OOM killer may be shooting >>> things down. When it happens, you will see something like "'killed process" >>> in system's syslog. >>> >>>Thanks, >>> +Vinod >>> >>> On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri < >>> write2kish...@gmail.com> wrote: >>> >>> Vinod, >>> >>> One more thing I observed is that, my Client which submits Application >>> Master one after another continuously also gets killed sometimes. So, it is >>> always any of the Java Processes that is getting killed. Does it indicate >>> some excessive memory usage by them or something like that, that is causing >>> them die? If so, how can we resolve this kind of issue? >>> >>> Thanks, >>> Kishore >>> >>> >>> On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri < >>> write2kish...@gmail.com> wrote: >>> >>>> No, I am running on 2 node cluster. >>>> >>>> >>>> O
Re: Yarn -- one of the daemons getting killed
That's good info. It is more than likely that it is the OOM killer. See http://stackoverflow.com/questions/726690/who-killed-my-process-and-why for example. Thanks, +Vinod On Dec 17, 2013, at 1:26 AM, Krishna Kishore Bonagiri wrote: > Hi Jeff, > > I have run the resource manager in the foreground without nohup and here > are the messages when it was killed, it says it is "Killed" but doesn't say > why! > > 13/12/17 03:14:54 INFO capacity.CapacityScheduler: Application > appattempt_1387266015651_0258_01 released container > container_1387266015651_0258_01_03 on node: host: isredeng:36576 > #containers=2 available=7936 used=256 with event: FINISHED > 13/12/17 03:14:54 INFO rmcontainer.RMContainerImpl: > container_1387266015651_0258_01_05 Container Transitioned from ACQUIRED > to RUNNING > Killed > > > Thanks, > Kishore > > > On Mon, Dec 16, 2013 at 11:10 PM, Jeff Stuckman wrote: > What if you open the daemons in a "screen" session rather than running them > in the background -- for example, run "yarn resourcemanager". Then you can > see exactly when they terminate, and hopefully why. > > From: Krishna Kishore Bonagiri > Sent: Monday, December 16, 2013 6:20 AM > To: user@hadoop.apache.org > Reply To: user@hadoop.apache.org > Subject: Re: Yarn -- one of the daemons getting killed > > Hi Vinod, > > Yes, I am running on Linux. > > I was actually searching for a corresponding message in /var/log/messages to > confirm that OOM killed my daemons, but could not find any corresponding > messages there! According to the following link, it looks like if it is a > memory issue, I should see a messages even if OOM is disabled, but I don't > see it. > > http://www.redhat.com/archives/taroon-list/2007-August/msg6.html > > And, is memory consumption more in case of two node cluster than a single > node one? Also, I see this problem only when I give "*" as the node name. > > One other thing I suspected was the allowed number of user processes, I > increased that to 31000 from 1024 but that also didn't help. > > Thanks, > Kishore > > > On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli > wrote: > Yes, that is what I suspect. That is why I asked if everything is on a single > node. If you are running linux, linux OOM killer may be shooting things down. > When it happens, you will see something like "'killed process" in system's > syslog. > > Thanks, > +Vinod > > On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri > wrote: > >> Vinod, >> >> One more thing I observed is that, my Client which submits Application >> Master one after another continuously also gets killed sometimes. So, it is >> always any of the Java Processes that is getting killed. Does it indicate >> some excessive memory usage by them or something like that, that is causing >> them die? If so, how can we resolve this kind of issue? >> >> Thanks, >> Kishore >> >> >> On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri >> wrote: >> No, I am running on 2 node cluster. >> >> >> On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli >> wrote: >> Is all of this on a single node? >> >> Thanks, >> +Vinod >> >> On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri >> wrote: >> >>> Hi, >>> I am running a small application on YARN (2.2.0) in a loop of 500 times, >>> and while doing so one of the daemons, node manager, resource manager, or >>> data node is getting killed (I mean disappearing) at a random point. I see >>> no information in the corresponding log files. How can I know why is it >>> happening so? >>> >>> And, one more observation is that, this is happening only when I am using >>> "*" for node name in the container requests, otherwise when I used a >>> specific node name, everything is fine. >>> >>> Thanks, >>> Kishore >> >> >> CONFIDENTIALITY NOTICE >> NOTICE: This message is intended for the use of the individual or entity to >> which it is addressed and may contain information that is confidential, >> privileged and exempt from disclosure under applicable law. If the reader of >> this message is not the intended recipient, you are hereby notified that any >> printing, copying, dissemination, distribution, disclosure or forwarding of >> this communication is strictly prohibited. If you have received this >> comm
Re: Yarn -- one of the daemons getting killed
Hi Jeff, I have run the resource manager in the foreground without nohup and here are the messages when it was killed, it says it is "Killed" but doesn't say why! 13/12/17 03:14:54 INFO capacity.CapacityScheduler: Application appattempt_1387266015651_0258_01 released container container_1387266015651_0258_01_03 on node: host: isredeng:36576 #containers=2 available=7936 used=256 with event: FINISHED 13/12/17 03:14:54 INFO rmcontainer.RMContainerImpl: container_1387266015651_0258_01_05 Container Transitioned from ACQUIRED to RUNNING Killed Thanks, Kishore On Mon, Dec 16, 2013 at 11:10 PM, Jeff Stuckman wrote: > What if you open the daemons in a "screen" session rather than running > them in the background -- for example, run "yarn resourcemanager". Then you > can see exactly when they terminate, and hopefully why. > >*From: *Krishna Kishore Bonagiri > *Sent: *Monday, December 16, 2013 6:20 AM > *To: *user@hadoop.apache.org > *Reply To: *user@hadoop.apache.org > *Subject: *Re: Yarn -- one of the daemons getting killed > > Hi Vinod, > > Yes, I am running on Linux. > > I was actually searching for a corresponding message in /var/log/messages > to confirm that OOM killed my daemons, but could not find any corresponding > messages there! According to the following link, it looks like if it is a > memory issue, I should see a messages even if OOM is disabled, but I don't > see it. > > http://www.redhat.com/archives/taroon-list/2007-August/msg6.html > >And, is memory consumption more in case of two node cluster than a > single node one? Also, I see this problem only when I give "*" as the node > name. > >One other thing I suspected was the allowed number of user processes, > I increased that to 31000 from 1024 but that also didn't help. > > Thanks, > Kishore > > > On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli < > vino...@hortonworks.com> wrote: > >> Yes, that is what I suspect. That is why I asked if everything is on a >> single node. If you are running linux, linux OOM killer may be shooting >> things down. When it happens, you will see something like "'killed process" >> in system's syslog. >> >>Thanks, >> +Vinod >> >> On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri < >> write2kish...@gmail.com> wrote: >> >> Vinod, >> >> One more thing I observed is that, my Client which submits Application >> Master one after another continuously also gets killed sometimes. So, it is >> always any of the Java Processes that is getting killed. Does it indicate >> some excessive memory usage by them or something like that, that is causing >> them die? If so, how can we resolve this kind of issue? >> >> Thanks, >> Kishore >> >> >> On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri < >> write2kish...@gmail.com> wrote: >> >>> No, I am running on 2 node cluster. >>> >>> >>> On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli < >>> vino...@hortonworks.com> wrote: >>> >>>> Is all of this on a single node? >>>> >>>> Thanks, >>>> +Vinod >>>> >>>> On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri < >>>> write2kish...@gmail.com> wrote: >>>> >>>> Hi, >>>> I am running a small application on YARN (2.2.0) in a loop of 500 >>>> times, and while doing so one of the daemons, node manager, resource >>>> manager, or data node is getting killed (I mean disappearing) at a random >>>> point. I see no information in the corresponding log files. How can I know >>>> why is it happening so? >>>> >>>> And, one more observation is that, this is happening only when I am >>>> using "*" for node name in the container requests, otherwise when I used a >>>> specific node name, everything is fine. >>>> >>>> Thanks, >>>> Kishore >>>> >>>> >>>> >>>> CONFIDENTIALITY NOTICE >>>> NOTICE: This message is intended for the use of the individual or >>>> entity to which it is addressed and may contain information that is >>>> confidential, privileged and exempt from disclosure under applicable law. >>>> If the reader of this message is not the intended recipient, you are hereby >>>> notified that any printing, copying, dissemination, distribution, >>>> disclosure or forwarding of this communication is strictly prohibited. If >>>> you have received this communication in error, please contact the sender >>>> immediately and delete it from your system. Thank You. >>> >>> >>> >> >> >> CONFIDENTIALITY NOTICE >> NOTICE: This message is intended for the use of the individual or entity >> to which it is addressed and may contain information that is confidential, >> privileged and exempt from disclosure under applicable law. If the reader >> of this message is not the intended recipient, you are hereby notified that >> any printing, copying, dissemination, distribution, disclosure or >> forwarding of this communication is strictly prohibited. If you have >> received this communication in error, please contact the sender immediately >> and delete it from your system. Thank You. >> > >
RE: Yarn -- one of the daemons getting killed
If it is not killed by OOM killer, maybe the JVM just did a core dump due to whatever reason. Search for core dump of process in the /var/log/messages, or core dump file in your system. From: stuck...@umd.edu To: user@hadoop.apache.org; user@hadoop.apache.org Subject: Re: Yarn -- one of the daemons getting killed Date: Mon, 16 Dec 2013 17:40:10 + What if you open the daemons in a "screen" session rather than running them in the background -- for example, run "yarn resourcemanager". Then you can see exactly when they terminate, and hopefully why. From: Krishna Kishore Bonagiri Sent: Monday, December 16, 2013 6:20 AM To: user@hadoop.apache.org Reply To: user@hadoop.apache.org Subject: Re: Yarn -- one of the daemons getting killed Hi Vinod, Yes, I am running on Linux. I was actually searching for a corresponding message in /var/log/messages to confirm that OOM killed my daemons, but could not find any corresponding messages there! According to the following link, it looks like if it is a memory issue, I should see a messages even if OOM is disabled, but I don't see it. http://www.redhat.com/archives/taroon-list/2007-August/msg6.html And, is memory consumption more in case of two node cluster than a single node one? Also, I see this problem only when I give "*" as the node name. One other thing I suspected was the allowed number of user processes, I increased that to 31000 from 1024 but that also didn't help. Thanks, Kishore On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli wrote: Yes, that is what I suspect. That is why I asked if everything is on a single node. If you are running linux, linux OOM killer may be shooting things down. When it happens, you will see something like "'killed process" in system's syslog. Thanks, +Vinod On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri wrote: Vinod, One more thing I observed is that, my Client which submits Application Master one after another continuously also gets killed sometimes. So, it is always any of the Java Processes that is getting killed. Does it indicate some excessive memory usage by them or something like that, that is causing them die? If so, how can we resolve this kind of issue? Thanks, Kishore On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri wrote: No, I am running on 2 node cluster. On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli wrote: Is all of this on a single node? Thanks, +Vinod On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri wrote: Hi, I am running a small application on YARN (2.2.0) in a loop of 500 times, and while doing so one of the daemons, node manager, resource manager, or data node is getting killed (I mean disappearing) at a random point. I see no information in the corresponding log files. How can I know why is it happening so? And, one more observation is that, this is happening only when I am using "*" for node name in the container requests, otherwise when I used a specific node name, everything is fine. Thanks, Kishore CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You. CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.
Re: Yarn -- one of the daemons getting killed
What if you open the daemons in a "screen" session rather than running them in the background -- for example, run "yarn resourcemanager". Then you can see exactly when they terminate, and hopefully why. From: Krishna Kishore Bonagiri Sent: Monday, December 16, 2013 6:20 AM To: user@hadoop.apache.org Reply To: user@hadoop.apache.org Subject: Re: Yarn -- one of the daemons getting killed Hi Vinod, Yes, I am running on Linux. I was actually searching for a corresponding message in /var/log/messages to confirm that OOM killed my daemons, but could not find any corresponding messages there! According to the following link, it looks like if it is a memory issue, I should see a messages even if OOM is disabled, but I don't see it. http://www.redhat.com/archives/taroon-list/2007-August/msg6.html And, is memory consumption more in case of two node cluster than a single node one? Also, I see this problem only when I give "*" as the node name. One other thing I suspected was the allowed number of user processes, I increased that to 31000 from 1024 but that also didn't help. Thanks, Kishore On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli mailto:vino...@hortonworks.com>> wrote: Yes, that is what I suspect. That is why I asked if everything is on a single node. If you are running linux, linux OOM killer may be shooting things down. When it happens, you will see something like "'killed process" in system's syslog. Thanks, +Vinod On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri mailto:write2kish...@gmail.com>> wrote: Vinod, One more thing I observed is that, my Client which submits Application Master one after another continuously also gets killed sometimes. So, it is always any of the Java Processes that is getting killed. Does it indicate some excessive memory usage by them or something like that, that is causing them die? If so, how can we resolve this kind of issue? Thanks, Kishore On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri mailto:write2kish...@gmail.com>> wrote: No, I am running on 2 node cluster. On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli mailto:vino...@hortonworks.com>> wrote: Is all of this on a single node? Thanks, +Vinod On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri mailto:write2kish...@gmail.com>> wrote: Hi, I am running a small application on YARN (2.2.0) in a loop of 500 times, and while doing so one of the daemons, node manager, resource manager, or data node is getting killed (I mean disappearing) at a random point. I see no information in the corresponding log files. How can I know why is it happening so? And, one more observation is that, this is happening only when I am using "*" for node name in the container requests, otherwise when I used a specific node name, everything is fine. Thanks, Kishore CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You. CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.
Re: Yarn -- one of the daemons getting killed
Hi Vinay, In the out files I could see nothing other than the output of ulimit -all . Do I need to enable any other kind of logging to get more information? Thanks, Kishore On Mon, Dec 16, 2013 at 5:41 PM, Vinayakumar B wrote: > Hi Krishna, > > > > Please check the out files as well for daemons. You may find something. > > > > > > Cheers, > > Vinayakumar B > > > > *From:* Krishna Kishore Bonagiri [mailto:write2kish...@gmail.com] > *Sent:* 16 December 2013 16:50 > *To:* user@hadoop.apache.org > *Subject:* Re: Yarn -- one of the daemons getting killed > > > > Hi Vinod, > > > > Yes, I am running on Linux. > > > > I was actually searching for a corresponding message in /var/log/messages > to confirm that OOM killed my daemons, but could not find any corresponding > messages there! According to the following link, it looks like if it is a > memory issue, I should see a messages even if OOM is disabled, but I don't > see it. > > > > http://www.redhat.com/archives/taroon-list/2007-August/msg6.html > > > > And, is memory consumption more in case of two node cluster than a > single node one? Also, I see this problem only when I give "*" as the node > name. > > > > One other thing I suspected was the allowed number of user processes, I > increased that to 31000 from 1024 but that also didn't help. > > > > Thanks, > > Kishore > > > > On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli < > vino...@hortonworks.com> wrote: > > Yes, that is what I suspect. That is why I asked if everything is on a > single node. If you are running linux, linux OOM killer may be shooting > things down. When it happens, you will see something like "'killed process" > in system's syslog. > > > > Thanks, > > +Vinod > > > > On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > > > Vinod, > > > > One more thing I observed is that, my Client which submits Application > Master one after another continuously also gets killed sometimes. So, it is > always any of the Java Processes that is getting killed. Does it indicate > some excessive memory usage by them or something like that, that is causing > them die? If so, how can we resolve this kind of issue? > > > > Thanks, > > Kishore > > > > On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > No, I am running on 2 node cluster. > > > > On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli < > vino...@hortonworks.com> wrote: > > Is all of this on a single node? > > > > Thanks, > > +Vinod > > > > On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > > > Hi, > > I am running a small application on YARN (2.2.0) in a loop of 500 times, > and while doing so one of the daemons, node manager, resource manager, or > data node is getting killed (I mean disappearing) at a random point. I see > no information in the corresponding log files. How can I know why is it > happening so? > > > > And, one more observation is that, this is happening only when I am using > "*" for node name in the container requests, otherwise when I used a > specific node name, everything is fine. > > > > Thanks, > > Kishore > > > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity > to which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader > of this message is not the intended recipient, you are hereby notified that > any printing, copying, dissemination, distribution, disclosure or > forwarding of this communication is strictly prohibited. If you have > received this communication in error, please contact the sender immediately > and delete it from your system. Thank You. > > > > > > > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity > to which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader > of this message is not the intended recipient, you are hereby notified that > any printing, copying, dissemination, distribution, disclosure or > forwarding of this communication is strictly prohibited. If you have > received this communication in error, please contact the sender immediately > and delete it from your system. Thank You. > > >
RE: Yarn -- one of the daemons getting killed
Hi Krishna, Please check the out files as well for daemons. You may find something. Cheers, Vinayakumar B From: Krishna Kishore Bonagiri [mailto:write2kish...@gmail.com] Sent: 16 December 2013 16:50 To: user@hadoop.apache.org Subject: Re: Yarn -- one of the daemons getting killed Hi Vinod, Yes, I am running on Linux. I was actually searching for a corresponding message in /var/log/messages to confirm that OOM killed my daemons, but could not find any corresponding messages there! According to the following link, it looks like if it is a memory issue, I should see a messages even if OOM is disabled, but I don't see it. http://www.redhat.com/archives/taroon-list/2007-August/msg6.html And, is memory consumption more in case of two node cluster than a single node one? Also, I see this problem only when I give "*" as the node name. One other thing I suspected was the allowed number of user processes, I increased that to 31000 from 1024 but that also didn't help. Thanks, Kishore On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli mailto:vino...@hortonworks.com>> wrote: Yes, that is what I suspect. That is why I asked if everything is on a single node. If you are running linux, linux OOM killer may be shooting things down. When it happens, you will see something like "'killed process" in system's syslog. Thanks, +Vinod On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri mailto:write2kish...@gmail.com>> wrote: Vinod, One more thing I observed is that, my Client which submits Application Master one after another continuously also gets killed sometimes. So, it is always any of the Java Processes that is getting killed. Does it indicate some excessive memory usage by them or something like that, that is causing them die? If so, how can we resolve this kind of issue? Thanks, Kishore On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri mailto:write2kish...@gmail.com>> wrote: No, I am running on 2 node cluster. On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli mailto:vino...@hortonworks.com>> wrote: Is all of this on a single node? Thanks, +Vinod On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri mailto:write2kish...@gmail.com>> wrote: Hi, I am running a small application on YARN (2.2.0) in a loop of 500 times, and while doing so one of the daemons, node manager, resource manager, or data node is getting killed (I mean disappearing) at a random point. I see no information in the corresponding log files. How can I know why is it happening so? And, one more observation is that, this is happening only when I am using "*" for node name in the container requests, otherwise when I used a specific node name, everything is fine. Thanks, Kishore CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You. CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.
Re: Yarn -- one of the daemons getting killed
Hi Vinod, Yes, I am running on Linux. I was actually searching for a corresponding message in /var/log/messages to confirm that OOM killed my daemons, but could not find any corresponding messages there! According to the following link, it looks like if it is a memory issue, I should see a messages even if OOM is disabled, but I don't see it. http://www.redhat.com/archives/taroon-list/2007-August/msg6.html And, is memory consumption more in case of two node cluster than a single node one? Also, I see this problem only when I give "*" as the node name. One other thing I suspected was the allowed number of user processes, I increased that to 31000 from 1024 but that also didn't help. Thanks, Kishore On Fri, Dec 13, 2013 at 11:51 PM, Vinod Kumar Vavilapalli < vino...@hortonworks.com> wrote: > Yes, that is what I suspect. That is why I asked if everything is on a > single node. If you are running linux, linux OOM killer may be shooting > things down. When it happens, you will see something like "'killed process" > in system's syslog. > > Thanks, > +Vinod > > On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > Vinod, > > One more thing I observed is that, my Client which submits Application > Master one after another continuously also gets killed sometimes. So, it is > always any of the Java Processes that is getting killed. Does it indicate > some excessive memory usage by them or something like that, that is causing > them die? If so, how can we resolve this kind of issue? > > Thanks, > Kishore > > > On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > >> No, I am running on 2 node cluster. >> >> >> On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli < >> vino...@hortonworks.com> wrote: >> >>> Is all of this on a single node? >>> >>> Thanks, >>> +Vinod >>> >>> On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri < >>> write2kish...@gmail.com> wrote: >>> >>> Hi, >>> I am running a small application on YARN (2.2.0) in a loop of 500 >>> times, and while doing so one of the daemons, node manager, resource >>> manager, or data node is getting killed (I mean disappearing) at a random >>> point. I see no information in the corresponding log files. How can I know >>> why is it happening so? >>> >>> And, one more observation is that, this is happening only when I am >>> using "*" for node name in the container requests, otherwise when I used a >>> specific node name, everything is fine. >>> >>> Thanks, >>> Kishore >>> >>> >>> >>> CONFIDENTIALITY NOTICE >>> NOTICE: This message is intended for the use of the individual or entity >>> to which it is addressed and may contain information that is confidential, >>> privileged and exempt from disclosure under applicable law. If the reader >>> of this message is not the intended recipient, you are hereby notified that >>> any printing, copying, dissemination, distribution, disclosure or >>> forwarding of this communication is strictly prohibited. If you have >>> received this communication in error, please contact the sender immediately >>> and delete it from your system. Thank You. >> >> >> > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity > to which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader > of this message is not the intended recipient, you are hereby notified that > any printing, copying, dissemination, distribution, disclosure or > forwarding of this communication is strictly prohibited. If you have > received this communication in error, please contact the sender immediately > and delete it from your system. Thank You. >
Re: Yarn -- one of the daemons getting killed
If you are interested, please read how we run into OOM-killer issue that was killing our TaskTrackers http://hakunamapdata.com/two-memory-related-issues-on-the-apache-hadoop-cluster/ (+ one issue related to heavy swapping). 2013/12/13 Vinod Kumar Vavilapalli > Yes, that is what I suspect. That is why I asked if everything is on a > single node. If you are running linux, linux OOM killer may be shooting > things down. When it happens, you will see something like "'killed process" > in system's syslog. > > Thanks, > +Vinod > > On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > Vinod, > > One more thing I observed is that, my Client which submits Application > Master one after another continuously also gets killed sometimes. So, it is > always any of the Java Processes that is getting killed. Does it indicate > some excessive memory usage by them or something like that, that is causing > them die? If so, how can we resolve this kind of issue? > > Thanks, > Kishore > > > On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > >> No, I am running on 2 node cluster. >> >> >> On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli < >> vino...@hortonworks.com> wrote: >> >>> Is all of this on a single node? >>> >>> Thanks, >>> +Vinod >>> >>> On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri < >>> write2kish...@gmail.com> wrote: >>> >>> Hi, >>> I am running a small application on YARN (2.2.0) in a loop of 500 >>> times, and while doing so one of the daemons, node manager, resource >>> manager, or data node is getting killed (I mean disappearing) at a random >>> point. I see no information in the corresponding log files. How can I know >>> why is it happening so? >>> >>> And, one more observation is that, this is happening only when I am >>> using "*" for node name in the container requests, otherwise when I used a >>> specific node name, everything is fine. >>> >>> Thanks, >>> Kishore >>> >>> >>> >>> CONFIDENTIALITY NOTICE >>> NOTICE: This message is intended for the use of the individual or entity >>> to which it is addressed and may contain information that is confidential, >>> privileged and exempt from disclosure under applicable law. If the reader >>> of this message is not the intended recipient, you are hereby notified that >>> any printing, copying, dissemination, distribution, disclosure or >>> forwarding of this communication is strictly prohibited. If you have >>> received this communication in error, please contact the sender immediately >>> and delete it from your system. Thank You. >> >> >> > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity > to which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader > of this message is not the intended recipient, you are hereby notified that > any printing, copying, dissemination, distribution, disclosure or > forwarding of this communication is strictly prohibited. If you have > received this communication in error, please contact the sender immediately > and delete it from your system. Thank You. >
Re: Yarn -- one of the daemons getting killed
Yes, that is what I suspect. That is why I asked if everything is on a single node. If you are running linux, linux OOM killer may be shooting things down. When it happens, you will see something like "'killed process" in system's syslog. Thanks, +Vinod On Dec 13, 2013, at 4:52 AM, Krishna Kishore Bonagiri wrote: > Vinod, > > One more thing I observed is that, my Client which submits Application > Master one after another continuously also gets killed sometimes. So, it is > always any of the Java Processes that is getting killed. Does it indicate > some excessive memory usage by them or something like that, that is causing > them die? If so, how can we resolve this kind of issue? > > Thanks, > Kishore > > > On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri > wrote: > No, I am running on 2 node cluster. > > > On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli > wrote: > Is all of this on a single node? > > Thanks, > +Vinod > > On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri > wrote: > >> Hi, >> I am running a small application on YARN (2.2.0) in a loop of 500 times, >> and while doing so one of the daemons, node manager, resource manager, or >> data node is getting killed (I mean disappearing) at a random point. I see >> no information in the corresponding log files. How can I know why is it >> happening so? >> >> And, one more observation is that, this is happening only when I am using >> "*" for node name in the container requests, otherwise when I used a >> specific node name, everything is fine. >> >> Thanks, >> Kishore > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity to > which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader of > this message is not the intended recipient, you are hereby notified that any > printing, copying, dissemination, distribution, disclosure or forwarding of > this communication is strictly prohibited. If you have received this > communication in error, please contact the sender immediately and delete it > from your system. Thank You. > > -- CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.
Re: Yarn -- one of the daemons getting killed
Vinod, One more thing I observed is that, my Client which submits Application Master one after another continuously also gets killed sometimes. So, it is always any of the Java Processes that is getting killed. Does it indicate some excessive memory usage by them or something like that, that is causing them die? If so, how can we resolve this kind of issue? Thanks, Kishore On Fri, Dec 13, 2013 at 10:16 AM, Krishna Kishore Bonagiri < write2kish...@gmail.com> wrote: > No, I am running on 2 node cluster. > > > On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli < > vino...@hortonworks.com> wrote: > >> Is all of this on a single node? >> >> Thanks, >> +Vinod >> >> On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri < >> write2kish...@gmail.com> wrote: >> >> Hi, >> I am running a small application on YARN (2.2.0) in a loop of 500 >> times, and while doing so one of the daemons, node manager, resource >> manager, or data node is getting killed (I mean disappearing) at a random >> point. I see no information in the corresponding log files. How can I know >> why is it happening so? >> >> And, one more observation is that, this is happening only when I am >> using "*" for node name in the container requests, otherwise when I used a >> specific node name, everything is fine. >> >> Thanks, >> Kishore >> >> >> >> CONFIDENTIALITY NOTICE >> NOTICE: This message is intended for the use of the individual or entity >> to which it is addressed and may contain information that is confidential, >> privileged and exempt from disclosure under applicable law. If the reader >> of this message is not the intended recipient, you are hereby notified that >> any printing, copying, dissemination, distribution, disclosure or >> forwarding of this communication is strictly prohibited. If you have >> received this communication in error, please contact the sender immediately >> and delete it from your system. Thank You. > > >
Re: Yarn -- one of the daemons getting killed
No, I am running on 2 node cluster. On Fri, Dec 13, 2013 at 1:52 AM, Vinod Kumar Vavilapalli < vino...@hortonworks.com> wrote: > Is all of this on a single node? > > Thanks, > +Vinod > > On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri < > write2kish...@gmail.com> wrote: > > Hi, > I am running a small application on YARN (2.2.0) in a loop of 500 times, > and while doing so one of the daemons, node manager, resource manager, or > data node is getting killed (I mean disappearing) at a random point. I see > no information in the corresponding log files. How can I know why is it > happening so? > > And, one more observation is that, this is happening only when I am using > "*" for node name in the container requests, otherwise when I used a > specific node name, everything is fine. > > Thanks, > Kishore > > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity > to which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader > of this message is not the intended recipient, you are hereby notified that > any printing, copying, dissemination, distribution, disclosure or > forwarding of this communication is strictly prohibited. If you have > received this communication in error, please contact the sender immediately > and delete it from your system. Thank You.
Re: Yarn -- one of the daemons getting killed
Is all of this on a single node? Thanks, +Vinod On Dec 12, 2013, at 3:26 AM, Krishna Kishore Bonagiri wrote: > Hi, > I am running a small application on YARN (2.2.0) in a loop of 500 times, > and while doing so one of the daemons, node manager, resource manager, or > data node is getting killed (I mean disappearing) at a random point. I see no > information in the corresponding log files. How can I know why is it > happening so? > > And, one more observation is that, this is happening only when I am using > "*" for node name in the container requests, otherwise when I used a specific > node name, everything is fine. > > Thanks, > Kishore -- CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.