Re: Standalone cluster not using multiple workers for single application
With the default configuration SparkTC won’t run on my cluster. The log has: 15/11/03 17:50:13 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources With the SparkUI Completed Applications app-20151103172824-0290 SparkTC 2147483647512.0 MB2015/11/03 17:28:24adaptive-testFINISHED2 s However if I set spark.cores.max=16 then it runs but it’ll only use one worker. Workers Worker IdAddressStateCoresMemory worker-20150920064814-10.248.0.102-33659 10.248.0.102:33659ALIVE4 (0 Used)28.0 GB (0.0 B Used) worker-20151012175609-10.248.0.242-37399 10.248.0.242:37399ALIVE4 (16 Used)28.0 GB (512.0 MB Used) worker-20151012181934-10.248.0.188-36573 10.248.0.188:36573ALIVE4 (4 Used)28.0 GB (28.0 GB Used) worker-20151030170514-10.248.0.218-45368 10.248.0.218:45368ALIVE4 (0 Used)28.0 GB (0.0 B Used) Running Applications Application IDNameCoresMemory per NodeSubmitted TimeUserStateDuration app-20151103174134-0292 (kill) SparkTC 16512.0 MB2015/11/03 17:41:34adaptive-testRUNNING6 s app-20151103172428-0289 (kill) ImmunoSeq 428.0 GB2015/11/03 17:24:28adaptive-testRUNNING17 min On 11/2/15, 9:17 PM, "Jean-Baptiste Onofré"wrote: >Hi Jeff, > >it may depend of your application code. > >To verify your setup and if your are able to scale on multiple worker, >you can try using the SparkTC example for instance (it should use all >workers). > >Regards >JB > >On 11/02/2015 08:56 PM, Jeff Jones wrote: >> I’ve got an a series of applications using a single standalone Spark >> cluster (v1.4.1). The cluster has 1 master and 4 workers (4 CPUs per >> worker node). I am using the start-slave.sh script to launch the worker >> process on each node and I can see the nodes were successfully >> registered using the SparkUI. When I launch one of my applications >> regardless of what I set spark.cores.max to when instantiating the >> SparkContext in the driver app I seem to get a single worker assigned to >> the application and all jobs that get run. For example, if I set >> spark.cores.max to 16 the SparkUI will show a single worker take the >> load with 4 (16 Used) in the Cores column. How do I get my jobs run >> across multiple nodes in the cluster? >> >> Here’s a snippet from the SparkUI (IP addresses removed for privacy) >> >> >> Workers >> >> Worker Id Address State Cores Memory >> worker-20150920064814-***-33659 ***:33659 ALIVE 4 (0 Used) 28.0 GB >> (0.0 B Used) >> worker-20151012175609-***37399 ***:37399 ALIVE 4 (16 Used) 28.0 GB >> (28.0 GB Used) >> worker-20151012181934-***-36573 ***:36573 ALIVE 4 (4 Used) 28.0 GB >> (28.0 GB Used) >> worker-20151030170514-***-45368 ***:45368 ALIVE 4 (0 Used) 28.0 GB >> (0.0 B Used) >> >> >> Running Applications >> >> Application ID Name Cores Memory per Node Submitted Time User >> State Duration >> app-20151102194733-0278 App1 16 28.0 GB 2015/11/02 19:47:33 *** >> RUNNING 2 s >> app-20151102164156-0274 App2 4 28.0 GB 2015/11/02 16:41:56 *** >> RUNNING 3.1 h >> >> Jeff >> >> >> This message (and any attachments) is intended only for the designated >> recipient(s). It >> may contain confidential or proprietary information, or have other >> limitations on use as >> indicated by the sender. If you are not a designated recipient, you may >> not review, use, >> copy or distribute this message. If you received this in error, please >> notify the sender by >> reply e-mail and delete this message. > >-- >Jean-Baptiste Onofré >jbono...@apache.org >http://blog.nanthrax.net >Talend - http://www.talend.com > >- >To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >For additional commands, e-mail: user-h...@spark.apache.org > This message (and any attachments) is intended only for the designated recipient(s). It may contain confidential or proprietary information, or have other limitations on use as indicated by the sender. If you are not a designated recipient, you may not review, use, copy or distribute this message. If you received this in error, please notify the sender by reply e-mail and delete this message.
Standalone cluster not using multiple workers for single application
I’ve got an a series of applications using a single standalone Spark cluster (v1.4.1). The cluster has 1 master and 4 workers (4 CPUs per worker node). I am using the start-slave.sh script to launch the worker process on each node and I can see the nodes were successfully registered using the SparkUI. When I launch one of my applications regardless of what I set spark.cores.max to when instantiating the SparkContext in the driver app I seem to get a single worker assigned to the application and all jobs that get run. For example, if I set spark.cores.max to 16 the SparkUI will show a single worker take the load with 4 (16 Used) in the Cores column. How do I get my jobs run across multiple nodes in the cluster? Here’s a snippet from the SparkUI (IP addresses removed for privacy) Workers Worker Id Address State Cores Memory worker-20150920064814-***-33659 ***:33659 ALIVE 4 (0 Used) 28.0 GB (0.0 B Used) worker-20151012175609-***37399 ***:37399 ALIVE 4 (16 Used) 28.0 GB (28.0 GB Used) worker-20151012181934-***-36573 ***:36573 ALIVE 4 (4 Used) 28.0 GB (28.0 GB Used) worker-20151030170514-***-45368 ***:45368 ALIVE 4 (0 Used) 28.0 GB (0.0 B Used) Running Applications Application ID NameCores Memory per Node Submitted Time UserState Duration app-20151102194733-0278 App116 28.0 GB 2015/11/02 19:47:33 *** RUNNING 2 s app-20151102164156-0274 App24 28.0 GB 2015/11/02 16:41:56 *** RUNNING 3.1 h Jeff This message (and any attachments) is intended only for the designated recipient(s). It may contain confidential or proprietary information, or have other limitations on use as indicated by the sender. If you are not a designated recipient, you may not review, use, copy or distribute this message. If you received this in error, please notify the sender by reply e-mail and delete this message.
Re: Standalone cluster not using multiple workers for single application
Hi Jeff, it may depend of your application code. To verify your setup and if your are able to scale on multiple worker, you can try using the SparkTC example for instance (it should use all workers). Regards JB On 11/02/2015 08:56 PM, Jeff Jones wrote: I’ve got an a series of applications using a single standalone Spark cluster (v1.4.1). The cluster has 1 master and 4 workers (4 CPUs per worker node). I am using the start-slave.sh script to launch the worker process on each node and I can see the nodes were successfully registered using the SparkUI. When I launch one of my applications regardless of what I set spark.cores.max to when instantiating the SparkContext in the driver app I seem to get a single worker assigned to the application and all jobs that get run. For example, if I set spark.cores.max to 16 the SparkUI will show a single worker take the load with 4 (16 Used) in the Cores column. How do I get my jobs run across multiple nodes in the cluster? Here’s a snippet from the SparkUI (IP addresses removed for privacy) Workers Worker Id Address State Cores Memory worker-20150920064814-***-33659 ***:33659 ALIVE 4 (0 Used) 28.0 GB (0.0 B Used) worker-20151012175609-***37399 ***:37399 ALIVE 4 (16 Used) 28.0 GB (28.0 GB Used) worker-20151012181934-***-36573 ***:36573 ALIVE 4 (4 Used) 28.0 GB (28.0 GB Used) worker-20151030170514-***-45368 ***:45368 ALIVE 4 (0 Used) 28.0 GB (0.0 B Used) Running Applications Application ID NameCores Memory per Node Submitted Time User State Duration app-20151102194733-0278 App116 28.0 GB 2015/11/02 19:47:33 *** RUNNING 2 s app-20151102164156-0274 App24 28.0 GB 2015/11/02 16:41:56 *** RUNNING 3.1 h Jeff This message (and any attachments) is intended only for the designated recipient(s). It may contain confidential or proprietary information, or have other limitations on use as indicated by the sender. If you are not a designated recipient, you may not review, use, copy or distribute this message. If you received this in error, please notify the sender by reply e-mail and delete this message. -- Jean-Baptiste Onofré jbono...@apache.org http://blog.nanthrax.net Talend - http://www.talend.com - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org