Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
Got it to work...thanks a lot for the help! I started a new cluster where Spark has Yarn as a dependency. I ran it with the script with local[2] and it worked (this same script did not work with Spark in standalone mode). A follow up question...I have seen this question posted around the internet

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
I thought I was running it in local mode as http://spark.apache.org/docs/1.1.1/submitting-applications.html says that if I don't include "--deploy-mode cluster" then it will run as local mode? I tried both of the scripts above and they gave the same result as the script I was running before. Also

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Akhil Das
You don't submit it like that :/ You use [*] things when you run the job in local mode, whereas here you are running it in stand alone cluster mode. You can try either of these: 1. /opt/cloudera/parcels/CDH-5.2.1-1.cdh5.2.10.12/lib/spark/bin/spark-submit --class SimpleApp --master spark://10.0.1

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
I tried submitting the application like this with 2 cores as you can see with the [2]. /opt/cloudera/parcels/CDH-5.2.1-1.cdh5.2.10.12/lib/spark/bin/spark-submit --class SimpleApp --master spark://10.0.1.230:7077[2] --jars $(echo /home/ec2-user/sparkApps/SimpleApp/lib/*.jar | tr ' ' ',') /home/ec2

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Akhil Das
How many cores are you allocated/seeing in the webui? (that usually runs on 8080, for cloudera i think its 18080). Most likely the job is being allocated 1 core (should be >= 2 cores) and that's why the count is never happening. Thanks Best Regards On Mon, Dec 29, 2014 at 2:22 PM, Suhas Shekar w

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
So it got rid of the logs, but the problem still persists that : a) The program never terminates (I have pasted all output after the Hello World statements below) b) I am not seeing the word count c) I tried adding [2] next to my 10.0.1.232:2181 looking at this post http://apache-spark-user-list

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Akhil Das
Now, Add these lines to get ride of those logs import org.apache.log4j.Logger import org.apache.log4j.Level Logger.getLogger("org").setLevel(Level.OFF) Logger.getLogger("akka").setLevel(Level.OFF) Thanks Best Regards On Mon, Dec 29, 2014 at 2:09 PM, Suhas Shekar wrote: > Hmmm

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
Hmmm..soo I added 1 (10,000) to jssc.awaitTermination , however it does not stop. When I am not pushing in any data it gives me this: 14/12/29 08:35:12 INFO ReceiverTracker: Stream 0 received 0 blocks 14/12/29 08:35:12 INFO JobScheduler: Added jobs for time 1419860112000 ms 14/12/29 08:35:14 I

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Akhil Das
If you want to stop the streaming after 10 seconds, then use ssc.awaitTermination(1). Make sure you push some data to kafka for the streaming to consume within the 10 seconds. Thanks Best Regards On Mon, Dec 29, 2014 at 1:53 PM, Suhas Shekar wrote: > I'm very close! So I added that and then

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
I'm very close! So I added that and then I added this: http://mvnrepository.com/artifact/org.apache.kafka/kafka-clients/0.8.2-beta and it seems as though the stream is working as it says Stream 0 received 1 or 2 blocks as I enter in messages on my kafka producer. However, the Receiver seems to kee

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Akhil Das
Add this jar in the dependency http://mvnrepository.com/artifact/com.yammer.metrics/metrics-core/2.2.0 Thanks Best Regards On Mon, Dec 29, 2014 at 1:31 PM, Suhas Shekar wrote: > Hello Akhil, > > I chanced my Kafka dependency to 2.10 (which is the version of kafka that > was on 10.0.1.232). I am

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-29 Thread Suhas Shekar
Hello Akhil, I chanced my Kafka dependency to 2.10 (which is the version of kafka that was on 10.0.1.232). I am getting a slightly different error, but at the same place as the previous error (pasted below). FYI, when I make these changes to the pom file, I do "mvn clean package" then cp the new

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-28 Thread Suhas Shekar
I made both versions 1.1.1 and I got the same error. I then tried making both 1.1.0 as that is the version of my Spark Core, but I got the same error. I noticed my Kafka dependency is for scala 2.9.2, while my spark streaming kafka dependency is 2.10.x...I will try changing that next, but don't th

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-28 Thread Akhil Das
Just looked at the pom file that you are using, why are you having different versions in it? org.apache.spark spark-streaming-kafka_2.10 *1.1.1* org.apache.spark spark-streaming_2.10 *1.0.2* ​can you make both the versions the same?​ Thanks Best Regards On Mon, Dec 29, 2014 at 12:44 PM, Su

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-28 Thread Suhas Shekar
1) Could you please clarify on what you mean by checking the Scala version is correct? In my pom.xml file it is 2.10.4 (which is the same as when I start spark-shell). 2) The spark master URL is definitely correct as I have run other apps with the same script that use Spark (like a word count with

Re: Setting up Simple Kafka Consumer via Spark Java app

2014-12-28 Thread Akhil Das
Make sure you verify the following: - Scala version : I think the correct version would be 2.10.x - SparkMasterURL: Be sure that you copied the one displayed on the webui's top left corner (running on port 8080) Thanks Best Regards On Mon, Dec 29, 2014 at 12:26 PM, suhshekar52 wrote: > Hello E