The code is pretty long. But the main idea is to consume from Kafka,
preprocess the data, and groupBy a field. I use mutliple DStream to add
parallelism to the consumer. It seems when the number of DStreams is large,
this happens often.
Thanks,
Bill
On Tue, Jul 22, 2014 at 11:13 PM, Akhil Das
Can you paste the piece of code?
Thanks
Best Regards
On Wed, Jul 23, 2014 at 1:22 AM, Bill Jay
wrote:
> Hi all,
>
> I am running a spark streaming job. The job hangs on one stage, which
> shows as follows:
>
> Details for Stage 4
> Summary Metrics No tasks have started yetTasksNo tasks have st
Hi all,
I am running a spark streaming job. The job hangs on one stage, which shows
as follows:
Details for Stage 4
Summary MetricsNo tasks have started yetTasksNo tasks have started yet
Does anyone have an idea on this?
Thanks!
Bill
Bill