Vivek,

https://spark.apache.org/docs/1.5.2/streaming-kafka-integration.html

The map is per partitions number of topics to consume. Is numThreads below 
equal to the number of partitions in your topic?

Regards,

Bryan Jeffrey

Sent from Outlook Mail for Windows 10 phone


From: vivek.meghanat...@wipro.com
Sent: Friday, December 25, 2015 2:18 PM
To: bryan.jeff...@gmail.com; user@spark.apache.org
Subject: Re: Spark Streaming + Kafka + scala job message read issue

Any help is highly appreciated, i am completely stuck here..


From: Vivek Meghanathan (WT01 - NEP)
Sent: Thursday, December 24, 2015 7:50 PM
To: Bryan; user@spark.apache.org
Subject: RE: Spark Streaming + Kafka + scala job message read issue 
 
We are using the older receiver based approach, the number of partitions is 1 
(we have a single node kafka) and we use single thread per topic still we have 
the problem. Please see the API we use. All 8 spark jobs use same group name – 
is that a problem?
 
val topicMap = topics.split(",").map((_, numThreads.toInt)).toMap  - Number of 
threads used here is 1
val searches = KafkaUtils.createStream(ssc, zkQuorum, group, topicMap).map(line 
=> parse(line._2).extract[Search])
 
 
Regards,
Vivek M
From: Bryan [mailto:bryan.jeff...@gmail.com] 
Sent: 24 December 2015 17:20
To: Vivek Meghanathan (WT01 - NEP) <vivek.meghanat...@wipro.com>; 
user@spark.apache.org
Subject: RE: Spark Streaming + Kafka + scala job message read issue
 
Are you using a direct stream consumer, or the older receiver based consumer? 
If the latter, do the number of partitions you’ve specified for your topic 
match the number of partitions in the topic on Kafka? 
 
That would be an possible cause – as you might receive all data from a given 
partition while missing data from other partitions.
 
Regards,
 
Bryan Jeffrey
 
Sent from Outlook Mail for Windows 10 phone
 

From: vivek.meghanat...@wipro.com
Sent: Thursday, December 24, 2015 5:22 AM
To: user@spark.apache.org
Subject: Spark Streaming + Kafka + scala job message read issue
 
Hi All,
 

We are using Bitnami Kafka 0.8.2 + spark 1.5.2 in Google cloud platform. Our 
spark streaming job(consumer) not receiving all the messages sent to the 
specific topic. It receives 1 out of ~50 messages(added log in the job stream 
and identified). We are not seeing any errors in the kafka logs. Unable to 
debug further from kafka layer. The console consumer shows the INPUT topic is 
received in the console. it is not reaching the spark-kafka integration stream. 
Any thoughts how to debug this issue. Another topic is working fine in same 
setup.
Again tried with spark 1.3.0, kafka 0.8.1.1 which is also has same issue. All 
these jobs are working fine in our local lab servers
Regards,
Vivek M
The information contained in this electronic message and any attachments to 
this message are intended for the exclusive use of the addressee(s) and may 
contain proprietary, confidential or privileged information. If you are not the 
intended recipient, you should not disseminate, distribute or copy this e-mail. 
Please notify the sender immediately and destroy all copies of this message and 
any attachments. WARNING: Computer viruses can be transmitted via email. The 
recipient should check this email and any attachments for the presence of 
viruses. The company accepts no liability for any damage caused by any virus 
transmitted by this email. www.wipro.com 
 
The information contained in this electronic message and any attachments to 
this message are intended for the exclusive use of the addressee(s) and may 
contain proprietary, confidential or privileged information. If you are not the 
intended recipient, you should not disseminate, distribute or copy this e-mail. 
Please notify the sender immediately and destroy all copies of this message and 
any attachments. WARNING: Computer viruses can be transmitted via email. The 
recipient should check this email and any attachments for the presence of 
viruses. The company accepts no liability for any damage caused by any virus 
transmitted by this email. www.wipro.com 

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to