Thank you bit1129,

>From looking at the web UI i can see 2 cores

Also looking at http://spark.apache.org/docs/1.2.1/configuration.html

But can't see obvious configuration for number of receivers can you help
please.


On Wed, Apr 1, 2015 at 9:39 AM, bit1...@163.com <bit1...@163.com> wrote:

> Please make sure that you have given more cores than Receiver numbers.
>
>
>
>
> *From:* James King <jakwebin...@gmail.com>
> *Date:* 2015-04-01 15:21
> *To:* user <user@spark.apache.org>
> *Subject:* Spark + Kafka
> I have a simple setup/runtime of Kafka and Sprak.
>
> I have a command line consumer displaying arrivals to Kafka topic. So i
> know messages are being received.
>
> But when I try to read from Kafka topic I get no messages, here are some
> logs below.
>
> I'm thinking there aren't enough threads. How do i check that.
>
> Thank you.
>
> 2015-04-01 08:56:50 INFO  JobScheduler:59 - Starting job streaming job
> 1427871410000 ms.0 from job set of time 1427871410000 ms
> 2015-04-01 08:56:50 INFO  JobScheduler:59 - Finished job streaming job
> 1427871410000 ms.0 from job set of time 1427871410000 ms
> 2015-04-01 08:56:50 INFO  JobScheduler:59 - Total delay: 0.002 s for time
> 1427871410000 ms (execution: 0.000 s)
> 2015-04-01 08:56:50 DEBUG JobGenerator:63 - Got event
> ClearMetadata(1427871410000 ms)
> 2015-04-01 08:56:50 DEBUG DStreamGraph:63 - Clearing metadata for time
> 1427871410000 ms
> 2015-04-01 08:56:50 DEBUG ForEachDStream:63 - Clearing references to old
> RDDs: []
> 2015-04-01 08:56:50 DEBUG ForEachDStream:63 - Unpersisting old RDDs:
> 2015-04-01 08:56:50 DEBUG ForEachDStream:63 - Cleared 0 RDDs that were
> older than 1427871405000 ms:
> 2015-04-01 08:56:50 DEBUG KafkaInputDStream:63 - Clearing references to
> old RDDs: [1427871405000 ms -> 8]
> 2015-04-01 08:56:50 DEBUG KafkaInputDStream:63 - Unpersisting old RDDs: 8
> 2015-04-01 08:56:50 INFO  BlockRDD:59 - Removing RDD 8 from persistence
> list
> 2015-04-01 08:56:50 DEBUG BlockManagerMasterActor:50 - [actor] received
> message RemoveRdd(8) from Actor[akka://sparkDriver/temp/$n]
> 2015-04-01 08:56:50 DEBUG BlockManagerSlaveActor:50 - [actor] received
> message RemoveRdd(8) from Actor[akka://sparkDriver/temp/$o]
> 2015-04-01 08:56:50 DEBUG BlockManagerMasterActor:56 - [actor] handled
> message (0.287257 ms) RemoveRdd(8) from Actor[akka://sparkDriver/temp/$n]
> 2015-04-01 08:56:50 DEBUG BlockManagerSlaveActor:63 - removing RDD 8
> 2015-04-01 08:56:50 INFO  BlockManager:59 - Removing RDD 8
> 2015-04-01 08:56:50 DEBUG BlockManagerSlaveActor:63 - Done removing RDD 8,
> response is 0
> 2015-04-01 08:56:50 DEBUG BlockManagerSlaveActor:56 - [actor] handled
> message (0.038047 ms) RemoveRdd(8) from Actor[akka://sparkDriver/temp/$o]
> 2015-04-01 08:56:50 INFO  KafkaInputDStream:59 - Removing blocks of RDD
> BlockRDD[8] at createStream at KafkaLogConsumer.java:53 of time
> 1427871410000 ms
> 2015-04-01 08:56:50 DEBUG KafkaInputDStream:63 - Cleared 1 RDDs that were
> older than 1427871405000 ms: 1427871405000 ms
> 2015-04-01 08:56:50 DEBUG DStreamGraph:63 - Cleared old metadata for time
> 1427871410000 ms
> 2015-04-01 08:56:50 INFO  ReceivedBlockTracker:59 - Deleting batches
> ArrayBuffer(1427871400000 ms)
> 2015-04-01 08:56:50 DEBUG BlockManagerSlaveActor:63 - Sent response: 0 to
> Actor[akka://sparkDriver/temp/$o]
> 2015-04-01 08:56:50 DEBUG SparkDeploySchedulerBackend:50 - [actor]
> received message ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:50 DEBUG TaskSchedulerImpl:63 - parentName: , name:
> TaskSet_0, runningTasks: 0
> 2015-04-01 08:56:50 DEBUG SparkDeploySchedulerBackend:56 - [actor] handled
> message (0.499181 ms) ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:50 WARN  TaskSchedulerImpl:71 - Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient resources
> 2015-04-01 08:56:51 DEBUG SparkDeploySchedulerBackend:50 - [actor]
> received message ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:51 DEBUG TaskSchedulerImpl:63 - parentName: , name:
> TaskSet_0, runningTasks: 0
> 2015-04-01 08:56:51 DEBUG SparkDeploySchedulerBackend:56 - [actor] handled
> message (0.886121 ms) ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:50 - [actor] received
> message ExecutorUpdated(0,EXITED,Some(Command exited with code 1),Some(1))
> from Actor[akka.tcp://sparkMaster@somesparkhost
> :7077/user/Master#336117298]
> 2015-04-01 08:56:52 INFO  AppClient$ClientActor:59 - Executor updated:
> app-20150401065621-0007/0 is now EXITED (Command exited with code 1)
> 2015-04-01 08:56:52 INFO  SparkDeploySchedulerBackend:59 - Executor
> app-20150401065621-0007/0 removed: Command exited with code 1
> 2015-04-01 08:56:52 DEBUG SparkDeploySchedulerBackend:50 - [actor]
> received message RemoveExecutor(0,Unknown executor exit code (1)) from
> Actor[akka://sparkDriver/temp/$p]
> 2015-04-01 08:56:52 ERROR SparkDeploySchedulerBackend:75 - Asked to remove
> non-existent executor 0
> 2015-04-01 08:56:52 DEBUG SparkDeploySchedulerBackend:56 - [actor] handled
> message (1.394052 ms) RemoveExecutor(0,Unknown executor exit code (1)) from
> Actor[akka://sparkDriver/temp/$p]
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:56 - [actor] handled
> message (6.653705 ms) ExecutorUpdated(0,EXITED,Some(Command exited with
> code 1),Some(1)) from Actor[akka.tcp://sparkMaster@somesparkhost
> :7077/user/Master#336117298]
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:50 - [actor] received
> message
> ExecutorAdded(1,worker-20150331133159-somesparkhost-49556,somesparkhost:49556,2,512)
> from Actor[akka.tcp://sparkMaster@somesparkhost
> :7077/user/Master#336117298]
> 2015-04-01 08:56:52 INFO  AppClient$ClientActor:59 - Executor added:
> app-20150401065621-0007/1 on worker-20150331133159-somesparkhost-49556
> (somesparkhost:49556) with 2 cores
> 2015-04-01 08:56:52 INFO  SparkDeploySchedulerBackend:59 - Granted
> executor ID app-20150401065621-0007/1 on hostPort somesparkhost:49556 with
> 2 cores, 512.0 MB RAM
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:56 - [actor] handled
> message (1.119731 ms)
> ExecutorAdded(1,worker-20150331133159-somesparkhost-49556,somesparkhost:49556,2,512)
> from Actor[akka.tcp://sparkMaster@somesparkhost
> :7077/user/Master#336117298]
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:50 - [actor] received
> message ExecutorUpdated(1,LOADING,None,None) from
> Actor[akka.tcp://sparkMaster@somesparkhost:7077/user/Master#336117298]
> 2015-04-01 08:56:52 INFO  AppClient$ClientActor:59 - Executor updated:
> app-20150401065621-0007/1 is now LOADING
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:56 - [actor] handled
> message (0.516301 ms) ExecutorUpdated(1,LOADING,None,None) from
> Actor[akka.tcp://sparkMaster@somesparkhost:7077/user/Master#336117298]
> 2015-04-01 08:56:52 DEBUG SparkDeploySchedulerBackend:50 - [actor]
> received message ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:52 DEBUG TaskSchedulerImpl:63 - parentName: , name:
> TaskSet_0, runningTasks: 0
> 2015-04-01 08:56:52 DEBUG SparkDeploySchedulerBackend:56 - [actor] handled
> message (0.652891 ms) ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:50 - [actor] received
> message ExecutorUpdated(1,RUNNING,None,None) from
> Actor[akka.tcp://sparkMaster@somesparkhost:7077/user/Master#336117298]
> 2015-04-01 08:56:52 INFO  AppClient$ClientActor:59 - Executor updated:
> app-20150401065621-0007/1 is now RUNNING
> 2015-04-01 08:56:52 DEBUG AppClient$ClientActor:56 - [actor] handled
> message (0.381614 ms) ExecutorUpdated(1,RUNNING,None,None) from
> Actor[akka.tcp://sparkMaster@somesparkhost:7077/user/Master#336117298]
> 2015-04-01 08:56:53 DEBUG SparkDeploySchedulerBackend:50 - [actor]
> received message ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:53 DEBUG TaskSchedulerImpl:63 - parentName: , name:
> TaskSet_0, runningTasks: 0
> 2015-04-01 08:56:53 DEBUG SparkDeploySchedulerBackend:56 - [actor] handled
> message (0.417759 ms) ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:54 DEBUG SparkDeploySchedulerBackend:50 - [actor]
> received message ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:54 DEBUG TaskSchedulerImpl:63 - parentName: , name:
> TaskSet_0, runningTasks: 0
> 2015-04-01 08:56:54 DEBUG SparkDeploySchedulerBackend:56 - [actor] handled
> message (1.426392 ms) ReviveOffers from
> Actor[akka://sparkDriver/user/CoarseGrainedScheduler#1727295119]
> 2015-04-01 08:56:55 DEBUG RecurringTimer:63 - Callback for JobGenerator
> called at time 1427871415000
> 2015-04-01 08:56:55 DEBUG JobGenerator:63 - Got event
> GenerateJobs(1427871415000 ms)
> 2015-04-01 08:56:55 DEBUG DStreamGraph:63 - Generating jobs for time
> 1427871415000 ms
> 2015-04-01 08:56:55 DEBUG KafkaInputDStream:63 - Time 1427871415000 ms is
> valid
> 2015-04-01 08:56:55 DEBUG DStreamGraph:63 - Generated 1 jobs for time
> 1427871415000 ms
> 2015-04-01 08:56:55 INFO  JobScheduler:59 - Added jobs for time
> 1427871415000 ms
> 2015-04-01 08:56:55 DEBUG JobGenerator:63 - Got event
> DoCheckpoint(1427871415000 ms)
> -------------------------------------------
> Time: 1427871415000 ms
> -------------------------------------------
>
>

Reply via email to