Hi

It could be due to the connectivity issue between the master and the slaves.

I have seen this issue occur for the following reasons.Are the slaves
visible in the Spark UI?And how much memory is allocated to the executors.

1. Syncing of configuration between Spark Master and Slaves.
2. Network connectivity issues between the master and slave.

Thanks
Arush

On Sat, Feb 14, 2015 at 3:07 PM, NORD SC <jan.algermis...@nordsc.com> wrote:

> Hi all,
>
> I am new to spark and seem to have hit a common newbie obstacle.
>
> I have a pretty simple setup and job but I am unable to get past this
> error when executing a job:
>
> "TaskSchedulerImpl: Initial job has not accepted any resources; check your
> cluster UI to ensure that workers are registered and have sufficient memory”
>
> I have so far gained a basic understanding of worker/executor/driver
> memory, but have run out of ideas what to try next - maybe someone has a
> clue.
>
>
> My setup:
>
> Three node standalone cluster with C* and spark on each node and the
> Datastax C*/Spark connector JAR placed on each node.
>
> On the master I have the slaves configured in conf/slaves and I am using
> sbin/start-all.sh to start the whole cluster.
>
> On each node I have this in conf/spark-defauls.conf
>
> spark.master                    spark://devpeng-db-cassandra-1:7077
> spark.eventLog.enabled   true
> spark.serializer                 org.apache.spark.serializer.KryoSerializer
>
> spark.executor.extraClassPath
> /opt/spark-cassandra-connector-assembly-1.2.0-alpha1.jar
>
> and this in conf/spart-env.sh
>
> SPARK_WORKER_MEMORY=6g
>
>
>
> My App looks like this
>
> object TestApp extends App {
>   val conf = new SparkConf(true).set("spark.cassandra.connection.host",
> "devpeng-db-cassandra-1.xxxxxxxx")
>   val sc = new SparkContext("spark://devpeng-db-cassandra-1:7077",
> "testApp", conf)
>   val rdd = sc.cassandraTable("test", "kv")
>   println(“Count: “ + String.valueOf(rdd.count) )
>   println(rdd.first)
> }
>
> Any kind of idea what to check next would help me at this point, I think.
>
> Jan
>
> Log of the application start:
>
> [info] Loading project definition from
> /Users/jan/projects/gkh/jump/workspace/gkh-spark-example/project
> [info] Set current project to csconnect (in build
> file:/Users/jan/projects/gkh/jump/workspace/gkh-spark-example/)
> [info] Compiling 1 Scala source to
> /Users/jan/projects/gkh/jump/workspace/gkh-spark-example/target/scala-2.10/classes...
> [info] Running jump.TestApp
> Using Spark's default log4j profile:
> org/apache/spark/log4j-defaults.properties
> 15/02/14 10:30:11 INFO SecurityManager: Changing view acls to: jan
> 15/02/14 10:30:11 INFO SecurityManager: Changing modify acls to: jan
> 15/02/14 10:30:11 INFO SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions: Set(jan); users
> with modify permissions: Set(jan)
> 15/02/14 10:30:11 INFO Slf4jLogger: Slf4jLogger started
> 15/02/14 10:30:11 INFO Remoting: Starting remoting
> 15/02/14 10:30:12 INFO Remoting: Remoting started; listening on addresses
> :[akka.tcp://sparkDriver@xxxxxx:58197]
> 15/02/14 10:30:12 INFO Utils: Successfully started service 'sparkDriver'
> on port 58197.
> 15/02/14 10:30:12 INFO SparkEnv: Registering MapOutputTracker
> 15/02/14 10:30:12 INFO SparkEnv: Registering BlockManagerMaster
> 15/02/14 10:30:12 INFO DiskBlockManager: Created local directory at
> /var/folders/vr/w3whx92d0356g5nj1p6s59gr0000gn/T/spark-local-20150214103012-5b53
> 15/02/14 10:30:12 INFO MemoryStore: MemoryStore started with capacity
> 530.3 MB
> 2015-02-14 10:30:12.304 java[24999:3b07] Unable to load realm info from
> SCDynamicStore
> 15/02/14 10:30:12 WARN NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 15/02/14 10:30:12 INFO HttpFileServer: HTTP File server directory is
> /var/folders/vr/w3whx92d0356g5nj1p6s59gr0000gn/T/spark-48459a22-c1ff-42d5-8b8e-cc89fe84933d
> 15/02/14 10:30:12 INFO HttpServer: Starting HTTP Server
> 15/02/14 10:30:12 INFO Utils: Successfully started service 'HTTP file
> server' on port 58198.
> 15/02/14 10:30:12 INFO Utils: Successfully started service 'SparkUI' on
> port 4040.
> 15/02/14 10:30:12 INFO SparkUI: Started SparkUI at http://xxxxxx:4040
> 15/02/14 10:30:12 INFO AppClient$ClientActor: Connecting to master
> spark://devpeng-db-cassandra-1:7077...
> 15/02/14 10:30:13 INFO SparkDeploySchedulerBackend: Connected to Spark
> cluster with app ID app-20150214103013-0001
> 15/02/14 10:30:13 INFO AppClient$ClientActor: Executor added:
> app-20150214103013-0001/0 on
> worker-20150214102534-devpeng-db-cassandra-2.devpengxxxx
> (devpeng-db-cassandra-2.devpeng.xxxxx:57563) with 8 cores
> 15/02/14 10:30:13 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20150214103013-0001/0 on hostPort
> devpeng-db-cassandra-2.devpeng.xxxx:57563 with 8 cores, 512.0 MB RAM
> 15/02/14 10:30:13 INFO AppClient$ClientActor: Executor added:
> app-20150214103013-0001/1 on
> worker-20150214102534-devpeng-db-cassandra-3.devpeng.xxxx-38773
> (devpeng-db-cassandra-3.devpeng.xxxxxx:38773) with 8 cores
> 15/02/14 10:30:13 INFO SparkDeploySchedulerBackend: Granted executor ID
> app-20150214103013-0001/1 on hostPort
> devpeng-db-cassandra-3.devpeng.xxxxxe:38773 with 8 cores, 512.0 MB RAM
> 15/02/14 10:30:13 INFO AppClient$ClientActor: Executor updated:
> app-20150214103013-0001/0 is now LOADING
> 15/02/14 10:30:13 INFO AppClient$ClientActor: Executor updated:
> app-20150214103013-0001/1 is now LOADING
> 15/02/14 10:30:13 INFO AppClient$ClientActor: Executor updated:
> app-20150214103013-0001/0 is now RUNNING
> 15/02/14 10:30:13 INFO AppClient$ClientActor: Executor updated:
> app-20150214103013-0001/1 is now RUNNING
> 15/02/14 10:30:13 INFO NettyBlockTransferService: Server created on 58200
> 15/02/14 10:30:13 INFO BlockManagerMaster: Trying to register BlockManager
> 15/02/14 10:30:13 INFO BlockManagerMasterActor: Registering block manager
> 192.168.2.103:58200 with 530.3 MB RAM, BlockManagerId(<driver>, xxxx,
> 58200)
> 15/02/14 10:30:13 INFO BlockManagerMaster: Registered BlockManager
> 15/02/14 10:30:13 INFO SparkDeploySchedulerBackend: SchedulerBackend is
> ready for scheduling beginning after reached minRegisteredResourcesRatio:
> 0.0
> 15/02/14 10:30:14 INFO Cluster: New Cassandra host
> devpeng-db-cassandra-1.devpeng.gkh-setu.de/xxxx:9042 added
> 15/02/14 10:30:14 INFO Cluster: New Cassandra host /xxx:9042 added
> 15/02/14 10:30:14 INFO Cluster: New Cassandra host xxxx:9042 added
> 15/02/14 10:30:14 INFO CassandraConnector: Connected to Cassandra cluster:
> GKHDevPeng
> 15/02/14 10:30:14 INFO LocalNodeFirstLoadBalancingPolicy: Adding host xxx
> (DC1)
> 15/02/14 10:30:14 INFO LocalNodeFirstLoadBalancingPolicy: Adding host xxx
> (DC1)
> 15/02/14 10:30:14 INFO LocalNodeFirstLoadBalancingPolicy: Adding host xxxx
> (DC1)
> 15/02/14 10:30:14 INFO LocalNodeFirstLoadBalancingPolicy: Adding host
> xxxxx (DC1)
> 15/02/14 10:30:14 INFO LocalNodeFirstLoadBalancingPolicy: Adding host
> xxxxx (DC1)
> 15/02/14 10:30:14 INFO LocalNodeFirstLoadBalancingPolicy: Adding host
> xxxxx (DC1)
> 15/02/14 10:30:15 INFO CassandraConnector: Disconnected from Cassandra
> cluster: GKHDevPeng
> 15/02/14 10:30:16 INFO SparkContext: Starting job: count at
> TestApp.scala:23
> 15/02/14 10:30:16 INFO DAGScheduler: Got job 0 (count at TestApp.scala:23)
> with 3 output partitions (allowLocal=false)
> 15/02/14 10:30:16 INFO DAGScheduler: Final stage: Stage 0(count at
> TestApp.scala:23)
> 15/02/14 10:30:16 INFO DAGScheduler: Parents of final stage: List()
> 15/02/14 10:30:16 INFO DAGScheduler: Missing parents: List()
> 15/02/14 10:30:16 INFO DAGScheduler: Submitting Stage 0 (CassandraRDD[0]
> at RDD at CassandraRDD.scala:49), which has no missing parents
> 15/02/14 10:30:16 INFO MemoryStore: ensureFreeSpace(4472) called with
> curMem=0, maxMem=556038881
> 15/02/14 10:30:16 INFO MemoryStore: Block broadcast_0 stored as values in
> memory (estimated size 4.4 KB, free 530.3 MB)
> 15/02/14 10:30:16 INFO MemoryStore: ensureFreeSpace(3082) called with
> curMem=4472, maxMem=556038881
> 15/02/14 10:30:16 INFO MemoryStore: Block broadcast_0_piece0 stored as
> bytes in memory (estimated size 3.0 KB, free 530.3 MB)
> 15/02/14 10:30:16 INFO BlockManagerInfo: Added broadcast_0_piece0 in
> memory on xxxxx  (size: 3.0 KB, free: 530.3 MB)
> 15/02/14 10:30:16 INFO BlockManagerMaster: Updated info of block
> broadcast_0_piece0
> 15/02/14 10:30:16 INFO SparkContext: Created broadcast 0 from broadcast at
> DAGScheduler.scala:838
> 15/02/14 10:30:16 INFO DAGScheduler: Submitting 3 missing tasks from Stage
> 0 (CassandraRDD[0] at RDD at CassandraRDD.scala:49)
> 15/02/14 10:30:16 INFO TaskSchedulerImpl: Adding task set 0.0 with 3 tasks
> 15/02/14 10:30:31 WARN TaskSchedulerImpl: Initial job has not accepted any
> resources; check your cluster UI to ensure that workers are registered and
> have sufficient memory
>
>
>
>
>
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>


-- 

[image: Sigmoid Analytics] <http://htmlsig.com/www.sigmoidanalytics.com>

*Arush Kharbanda* || Technical Teamlead

ar...@sigmoidanalytics.com || www.sigmoidanalytics.com

Reply via email to