RE: problem with start-slaves.sh
I also didn’t realize I was trying to bring up the 2ndNameNode as a slave.. that might be an issue as well.. Thanks, From: Yana Kadiyska [mailto:yana.kadiy...@gmail.com] Sent: Thursday, October 30, 2014 11:27 AM To: Pagliari, Roberto Cc: user@spark.apache.org Subject: Re: problem with start-slaves.sh Roberto, I don't think shark is an issue -- I have shark server running on a node that also acts as a worker. What you can do is turn off shark server, just run start-all to start your spark cluster. then you can try bin/spark-shell --master and see if you can successfully run some "hello world" stuff. This will verify you have a working Spark cluster. Shark is just an application on top of it, so I can't imagine that's what's causing interference. But stopping it is the simplest way to check. On Wed, Oct 29, 2014 at 10:54 PM, Pagliari, Roberto mailto:rpagli...@appcomsci.com>> wrote: hi Yana, in my case I did not start any spark worker. However, shark was definitely running. Do you think that might be a problem? I will take a look Thank you, From: Yana Kadiyska [yana.kadiy...@gmail.com<mailto:yana.kadiy...@gmail.com>] Sent: Wednesday, October 29, 2014 9:45 AM To: Pagliari, Roberto Cc: user@spark.apache.org<mailto:user@spark.apache.org> Subject: Re: problem with start-slaves.sh I see this when I start a worker and then try to start it again forgetting it's already running (I don't use start-slaves, I start the slaves individually with start-slave.sh). All this is telling you is that there is already a running process on that machine. You can see it if you do a ps -aef|grep worker you can look on the spark UI and see if your master shows this machine as connected to it already. If it doesn't, you might want to kill the worker process and restart it. On Tue, Oct 28, 2014 at 4:32 PM, Pagliari, Roberto mailto:rpagli...@appcomsci.com>> wrote: I ran sbin/start-master.sh followed by sbin/start-slaves.sh (I build with PHive option to be able to interface with hive) I’m getting this ip_address: org.apache.spark.deploy.worker.Worker running as process . Stop it first. Am I doing something wrong? In my specific case, shark+hive is running on the nodes. Does that interfere with spark? Thank you,
Re: problem with start-slaves.sh
Roberto, I don't think shark is an issue -- I have shark server running on a node that also acts as a worker. What you can do is turn off shark server, just run start-all to start your spark cluster. then you can try bin/spark-shell --master and see if you can successfully run some "hello world" stuff. This will verify you have a working Spark cluster. Shark is just an application on top of it, so I can't imagine that's what's causing interference. But stopping it is the simplest way to check. On Wed, Oct 29, 2014 at 10:54 PM, Pagliari, Roberto wrote: > hi Yana, > in my case I did not start any spark worker. However, shark was definitely > running. Do you think that might be a problem? > > I will take a look > > Thank you, > > -- > *From:* Yana Kadiyska [yana.kadiy...@gmail.com] > *Sent:* Wednesday, October 29, 2014 9:45 AM > *To:* Pagliari, Roberto > *Cc:* user@spark.apache.org > *Subject:* Re: problem with start-slaves.sh > > I see this when I start a worker and then try to start it again > forgetting it's already running (I don't use start-slaves, I start the > slaves individually with start-slave.sh). All this is telling you is that > there is already a running process on that machine. You can see it if you > do a ps -aef|grep worker > > you can look on the spark UI and see if your master shows this machine > as connected to it already. If it doesn't, you might want to kill the > worker process and restart it. > > On Tue, Oct 28, 2014 at 4:32 PM, Pagliari, Roberto < > rpagli...@appcomsci.com> wrote: > >> I ran sbin/start-master.sh followed by sbin/start-slaves.sh (I build >> with PHive option to be able to interface with hive) >> >> >> >> I’m getting this >> >> >> >> ip_address: org.apache.spark.deploy.worker.Worker running as process >> . Stop it first. >> >> >> >> Am I doing something wrong? In my specific case, shark+hive is running on >> the nodes. Does that interfere with spark? >> >> >> >> Thank you, >> > >
RE: problem with start-slaves.sh
hi Yana, in my case I did not start any spark worker. However, shark was definitely running. Do you think that might be a problem? I will take a look Thank you, From: Yana Kadiyska [yana.kadiy...@gmail.com] Sent: Wednesday, October 29, 2014 9:45 AM To: Pagliari, Roberto Cc: user@spark.apache.org Subject: Re: problem with start-slaves.sh I see this when I start a worker and then try to start it again forgetting it's already running (I don't use start-slaves, I start the slaves individually with start-slave.sh). All this is telling you is that there is already a running process on that machine. You can see it if you do a ps -aef|grep worker you can look on the spark UI and see if your master shows this machine as connected to it already. If it doesn't, you might want to kill the worker process and restart it. On Tue, Oct 28, 2014 at 4:32 PM, Pagliari, Roberto mailto:rpagli...@appcomsci.com>> wrote: I ran sbin/start-master.sh followed by sbin/start-slaves.sh (I build with PHive option to be able to interface with hive) I’m getting this ip_address: org.apache.spark.deploy.worker.Worker running as process . Stop it first. Am I doing something wrong? In my specific case, shark+hive is running on the nodes. Does that interfere with spark? Thank you,
Re: problem with start-slaves.sh
I see this when I start a worker and then try to start it again forgetting it's already running (I don't use start-slaves, I start the slaves individually with start-slave.sh). All this is telling you is that there is already a running process on that machine. You can see it if you do a ps -aef|grep worker you can look on the spark UI and see if your master shows this machine as connected to it already. If it doesn't, you might want to kill the worker process and restart it. On Tue, Oct 28, 2014 at 4:32 PM, Pagliari, Roberto wrote: > I ran sbin/start-master.sh followed by sbin/start-slaves.sh (I build with > PHive option to be able to interface with hive) > > > > I’m getting this > > > > ip_address: org.apache.spark.deploy.worker.Worker running as process . > Stop it first. > > > > Am I doing something wrong? In my specific case, shark+hive is running on > the nodes. Does that interfere with spark? > > > > Thank you, >