Hi, can somebody suggest me the way to reduce quantity of task?

2015-06-15 18:26 GMT+02:00 Serega Sheypak <serega.shey...@gmail.com>:

> Hi, I'm running spark sql against Cassandra table. I have 3 C* nodes, Each
> of them has spark worker.
> The problem is that spark runs 869 task to read 3 lines: select bar from
> foo.
> I've tried these properties:
>
> #try to avoid 769 tasks per dummy select foo from bar qeury
> spark.cassandra.input.split.size_in_mb=32mb
> spark.cassandra.input.fetch.size_in_rows=1000
> spark.cassandra.input.split.size=10000
>
> but it doesn't help.
>
> Here are  mean metrics for the job :
> input1= 8388608.0 TB
> input2 = -320 B
> input3 = -400 B
>
> I'm confused with input, there are only 3 rows in C* table.
> Definitely, I don't have 8388608.0 TB of data :)
>
>
>
>

Reply via email to