Added foreach as follows. Still don't see any output on my console. Would this go to the worker logs as Jerry indicated?
JavaPairReceiverInputDStream tweets = KafkaUtils.createStream(ssc, "<mymachine>:2181", "1", map); JavaDStream<String> statuses = tweets.map( new Function<String, String>() { public String call(String status) { return status; } } ); statuses.foreach(new Function<JavaRDD<String>, Void>() { @Override public Void call(JavaRDD<String> stringJavaRDD) throws Exception { for (String str: stringJavaRDD.take(10)) { System.out.println("Message: " + str); } return null; } }); On Tue, Nov 4, 2014 at 10:32 PM, Shao, Saisai <saisai.s...@intel.com> wrote: > If you’re running on a standalone mode, the log is under > <SPAR_HOME>/work/ directory. I’m not sure for yarn or mesos, you can check > the document of Spark to see the details. > > > > Thanks > > Jerry > > > > *From:* Something Something [mailto:mailinglist...@gmail.com] > *Sent:* Wednesday, November 05, 2014 2:28 PM > *To:* Shao, Saisai > *Cc:* user@spark.apache.org > *Subject:* Re: Kafka Consumer in Spark Streaming > > > > The Kafka broker definitely has messages coming in. But your #2 point is > valid. Needless to say I am a newbie to Spark. I can't figure out where > the 'executor' logs would be. How would I find them? > > All I see printed on my screen is this: > > 14/11/04 22:21:23 INFO Slf4jLogger: Slf4jLogger started > 14/11/04 22:21:23 INFO Remoting: Starting remoting > 14/11/04 22:21:24 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://spark@mymachie:60743] > 14/11/04 22:21:24 INFO Remoting: Remoting now listens on addresses: > [akka.tcp://spark@mymachine:60743] > 14/11/04 22:21:24 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 14/11/04 22:21:24 INFO JniBasedUnixGroupsMappingWithFallback: Falling back > to shell based > ------------------------------------------- > Time: 1415168520000 ms > ------------------------------------------- > ------------------------------------------- > Time: 1415168520000 ms > ------------------------------------------- > > Keeps repeating this... > > > > On Tue, Nov 4, 2014 at 10:14 PM, Shao, Saisai <saisai.s...@intel.com> > wrote: > > Hi, would you mind describing your problem a little more specific. > > > > 1. Is the Kafka broker currently has no data feed in? > > 2. This code will print the lines, but not in the driver side, the > code is running in the executor side, so you can check the log in worker > dir to see if there’s any printing logs under this folder. > > 3. Did you see any exceptions when running the app, this will help > to define the problem. > > > > Thanks > > Jerry > > > > *From:* Something Something [mailto:mailinglist...@gmail.com] > *Sent:* Wednesday, November 05, 2014 1:57 PM > *To:* user@spark.apache.org > *Subject:* Kafka Consumer in Spark Streaming > > > > I've following code in my program. I don't get any error, but it's not > consuming the messages either. Shouldn't the following code print the line > in the 'call' method? What am I missing? > > Please help. Thanks. > > > > JavaStreamingContext ssc = new JavaStreamingContext(sparkConf, new > Duration(60 * 1 * 1000)); > > JavaPairReceiverInputDStream tweets = KafkaUtils.createStream(ssc, > "<machine>:2181", "1", map); > > JavaDStream<String> statuses = tweets.map( > new Function<String, String>() { > public String call(String status) { > System.out.println(status); > return status; > } > } > ); > > >