I have noticed that the UI takes some time to reflect the requested changes. Is 
that the issue ? Have you tried waiting for a few minutes after killing the 
spark job from terminal ?

Regards,
Sudhanshu

Kind Regards,
Sudhanshu

On 23 Nov 2015, at 1:43 a.m., Ted Yu <yuzhih...@gmail.com> wrote:

>> If you ask about trapping the SIGKILL signal in your script, see the 
>> following:
>> 
>> http://linuxcommand.org/wss0160.php
>> 
>> Cheers
>  
>>> On Fri, Nov 20, 2015 at 10:02 PM, Vikram Kone <vikramk...@gmail.com> wrote:
>>> I tried adding shutdown hook to my code but it didn't help. Still same issue
>>> 
>>> 
>>>> On Fri, Nov 20, 2015 at 7:08 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>> Which Spark release are you using ?
>>>> 
>>>> Can you pastebin the stack trace of the process running on your machine ?
>>>> 
>>>> Thanks
>>>> 
>>>>> On Nov 20, 2015, at 6:46 PM, Vikram Kone <vikramk...@gmail.com> wrote:
>>>>> 
>>>>> Hi,
>>>>> I'm seeing a strange problem. I have a spark cluster in standalone mode. 
>>>>> I submit spark jobs from a remote node as follows from the terminal
>>>>> 
>>>>> spark-submit --master spark://10.1.40.18:7077  --class com.test.Ping 
>>>>> spark-jobs.jar
>>>>> 
>>>>> when the app is running , when I press ctrl-C on the console terminal, 
>>>>> then the process is killed and so is the app in the spark master UI. When 
>>>>> I go to spark master ui, i see that this app is in state Killed under 
>>>>> Completed applications, which is what I expected to see.
>>>>> 
>>>>> Now, I created a shell script as follows to do the same
>>>>> 
>>>>> #!/bin/bash
>>>>> spark-submit --master spark://10.1.40.18:7077  --class com.test.Ping 
>>>>> spark-jobs.jar
>>>>> echo $! > my.pid
>>>>> 
>>>>> When I execute the shell script from terminal, as follows
>>>>> 
>>>>> $> bash myscript.sh
>>>>> 
>>>>> The application is submitted correctly to spark master and I can see it 
>>>>> as one of the running apps in teh spark master ui. But when I kill the 
>>>>> process in my terminal as follows
>>>>> 
>>>>> $> ps kill $(cat my.pid)
>>>>> 
>>>>> I see that the process is killed on my machine but the spark appliation 
>>>>> is still running in spark master! It doesn't get killed.
>>>>> 
>>>>> I noticed one more thing that, when I launch the spark job via shell 
>>>>> script and kill the application from spark master UI by clicking on 
>>>>> "kill" next to the running application, it gets killed in spark ui but I 
>>>>> still see the process running in my machine. 
>>>>> 
>>>>> In both cases, I would expect the remote spark app to be killed and my 
>>>>> local process to be killed.
>>>>> 
>>>>> Why is this happening? and how can I kill a spark app from the terminal 
>>>>> launced via shell script w.o going to the spark master UI?
>>>>> 
>>>>> I want to launch the spark app via script and log the pid so i can 
>>>>> monitor it remotely
>>>>> 
>>>>> thanks for the help
> 

Reply via email to