Re: Standalone cluster not using multiple workers for single application

2015-11-03 Thread Jeff Jones
With the default configuration SparkTC won’t run on my cluster. The log has:

15/11/03 17:50:13 WARN TaskSchedulerImpl: Initial job has not accepted any 
resources; check your cluster UI to ensure that workers are registered and have 
sufficient resources

With the SparkUI Completed Applications
app-20151103172824-0290 SparkTC 2147483647512.0 MB2015/11/03 
17:28:24adaptive-testFINISHED2 s


However if I set spark.cores.max=16 then it runs but it’ll only use one worker.

Workers
Worker IdAddressStateCoresMemory
worker-20150920064814-10.248.0.102-33659 10.248.0.102:33659ALIVE4 (0 Used)28.0 
GB (0.0 B Used)
worker-20151012175609-10.248.0.242-37399 10.248.0.242:37399ALIVE4 (16 Used)28.0 
GB (512.0 MB Used)
worker-20151012181934-10.248.0.188-36573 10.248.0.188:36573ALIVE4 (4 Used)28.0 
GB (28.0 GB Used)
worker-20151030170514-10.248.0.218-45368 10.248.0.218:45368ALIVE4 (0 Used)28.0 
GB (0.0 B Used)



Running Applications
Application IDNameCoresMemory per NodeSubmitted TimeUserStateDuration
app-20151103174134-0292 (kill) SparkTC 16512.0 MB2015/11/03 
17:41:34adaptive-testRUNNING6 s
app-20151103172428-0289 (kill) ImmunoSeq 428.0 GB2015/11/03 
17:24:28adaptive-testRUNNING17 min







On 11/2/15, 9:17 PM, "Jean-Baptiste Onofré"  wrote:

>Hi Jeff,
>
>it may depend of your application code.
>
>To verify your setup and if your are able to scale on multiple worker,
>you can try using the SparkTC example for instance (it should use all
>workers).
>
>Regards
>JB
>
>On 11/02/2015 08:56 PM, Jeff Jones wrote:
>> I’ve got an a series of applications using a single standalone Spark
>> cluster (v1.4.1).  The cluster has 1 master and 4 workers (4 CPUs per
>> worker node). I am using the start-slave.sh script to launch the worker
>> process on each node and I can see the nodes were successfully
>> registered using the SparkUI.  When I launch one of my applications
>> regardless of what I set spark.cores.max to when instantiating the
>> SparkContext in the driver app I seem to get a single worker assigned to
>> the application and all jobs that get run.  For example, if I set
>> spark.cores.max to 16 the SparkUI will show a single worker take the
>> load with 4 (16 Used) in the Cores column.  How do I get my jobs run
>> across multiple nodes in the cluster?
>>
>> Here’s a snippet from the SparkUI (IP addresses removed for privacy)
>>
>>
>> Workers
>>
>> Worker Id Address State Cores Memory
>> worker-20150920064814-***-33659 ***:33659 ALIVE 4 (0 Used) 28.0 GB
>> (0.0 B Used)
>> worker-20151012175609-***37399 ***:37399 ALIVE 4 (16 Used) 28.0 GB
>> (28.0 GB Used)
>> worker-20151012181934-***-36573 ***:36573 ALIVE 4 (4 Used) 28.0 GB
>> (28.0 GB Used)
>> worker-20151030170514-***-45368 ***:45368 ALIVE 4 (0 Used) 28.0 GB
>> (0.0 B Used)
>>
>>
>> Running Applications
>>
>> Application ID Name Cores Memory per Node Submitted Time User
>> State Duration
>> app-20151102194733-0278 App1 16 28.0 GB 2015/11/02 19:47:33 ***
>> RUNNING 2 s
>> app-20151102164156-0274 App2 4 28.0 GB 2015/11/02 16:41:56 ***
>> RUNNING 3.1 h
>>
>> Jeff
>>
>>
>> This message (and any attachments) is intended only for the designated
>> recipient(s). It
>> may contain confidential or proprietary information, or have other
>> limitations on use as
>> indicated by the sender. If you are not a designated recipient, you may
>> not review, use,
>> copy or distribute this message. If you received this in error, please
>> notify the sender by
>> reply e-mail and delete this message.
>
>--
>Jean-Baptiste Onofré
>jbono...@apache.org
>http://blog.nanthrax.net
>Talend - http://www.talend.com
>
>-
>To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>For additional commands, e-mail: user-h...@spark.apache.org
>


This message (and any attachments) is intended only for the designated 
recipient(s). It
may contain confidential or proprietary information, or have other limitations 
on use as
indicated by the sender. If you are not a designated recipient, you may not 
review, use,
copy or distribute this message. If you received this in error, please notify 
the sender by
reply e-mail and delete this message.


Standalone cluster not using multiple workers for single application

2015-11-02 Thread Jeff Jones
I’ve got an a series of applications using a single standalone Spark cluster 
(v1.4.1).  The cluster has 1 master and 4 workers (4 CPUs per worker node). I 
am using the start-slave.sh script to launch the worker process on each node 
and I can see the nodes were successfully registered using the SparkUI.  When I 
launch one of my applications regardless of what I set spark.cores.max to when 
instantiating the SparkContext in the driver app I seem to get a single worker 
assigned to the application and all jobs that get run.  For example, if I set 
spark.cores.max to 16 the SparkUI will show a single worker take the load with 
4 (16 Used) in the Cores column.  How do I get my jobs run across multiple 
nodes in the cluster?

Here’s a snippet from the SparkUI (IP addresses removed for privacy)

Workers
Worker Id   Address State   Cores   Memory
worker-20150920064814-***-33659 ***:33659   ALIVE   4 (0 Used)  28.0 GB 
(0.0 B Used)
worker-20151012175609-***37399  ***:37399   ALIVE   4 (16 Used) 28.0 GB 
(28.0 GB Used)
worker-20151012181934-***-36573 ***:36573   ALIVE   4 (4 Used)  28.0 GB 
(28.0 GB Used)
worker-20151030170514-***-45368 ***:45368   ALIVE   4 (0 Used)  28.0 GB 
(0.0 B Used)
Running Applications
Application ID  NameCores   Memory per Node Submitted Time  UserState   
Duration
app-20151102194733-0278 App116  28.0 GB 2015/11/02 19:47:33 *** 
RUNNING 2 s
app-20151102164156-0274 App24   28.0 GB 2015/11/02 16:41:56 *** 
RUNNING 3.1 h
Jeff


This message (and any attachments) is intended only for the designated 
recipient(s). It
may contain confidential or proprietary information, or have other limitations 
on use as
indicated by the sender. If you are not a designated recipient, you may not 
review, use,
copy or distribute this message. If you received this in error, please notify 
the sender by
reply e-mail and delete this message.


Re: Standalone cluster not using multiple workers for single application

2015-11-02 Thread Jean-Baptiste Onofré

Hi Jeff,

it may depend of your application code.

To verify your setup and if your are able to scale on multiple worker, 
you can try using the SparkTC example for instance (it should use all 
workers).


Regards
JB

On 11/02/2015 08:56 PM, Jeff Jones wrote:

I’ve got an a series of applications using a single standalone Spark
cluster (v1.4.1).  The cluster has 1 master and 4 workers (4 CPUs per
worker node). I am using the start-slave.sh script to launch the worker
process on each node and I can see the nodes were successfully
registered using the SparkUI.  When I launch one of my applications
regardless of what I set spark.cores.max to when instantiating the
SparkContext in the driver app I seem to get a single worker assigned to
the application and all jobs that get run.  For example, if I set
spark.cores.max to 16 the SparkUI will show a single worker take the
load with 4 (16 Used) in the Cores column.  How do I get my jobs run
across multiple nodes in the cluster?

Here’s a snippet from the SparkUI (IP addresses removed for privacy)


Workers

Worker Id   Address State   Cores   Memory
worker-20150920064814-***-33659 ***:33659   ALIVE   4 (0 Used)  
28.0 GB
(0.0 B Used)
worker-20151012175609-***37399  ***:37399   ALIVE   4 (16 Used) 28.0 GB
(28.0 GB Used)
worker-20151012181934-***-36573 ***:36573   ALIVE   4 (4 Used)  
28.0 GB
(28.0 GB Used)
worker-20151030170514-***-45368 ***:45368   ALIVE   4 (0 Used)  
28.0 GB
(0.0 B Used)


Running Applications

Application ID  NameCores   Memory per Node Submitted Time  User
State   Duration
app-20151102194733-0278 App116  28.0 GB 2015/11/02 
19:47:33 ***
RUNNING 2 s
app-20151102164156-0274 App24   28.0 GB 2015/11/02 
16:41:56 ***
RUNNING 3.1 h

Jeff


This message (and any attachments) is intended only for the designated
recipient(s). It
may contain confidential or proprietary information, or have other
limitations on use as
indicated by the sender. If you are not a designated recipient, you may
not review, use,
copy or distribute this message. If you received this in error, please
notify the sender by
reply e-mail and delete this message.


--
Jean-Baptiste Onofré
jbono...@apache.org
http://blog.nanthrax.net
Talend - http://www.talend.com

-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org