Hello,

How feasible would it be to have spark speculatively increase the number of
partitions when there is spare capacity in the system? We want to do this
to increase to decrease application runtime. Initially, we will assume that
function calls of the same type will have the same runtime (e.g. all maps
take equal time) and that the runtime will scale linearly with the number
of workers. If a numPartitions value is specified, we may increase beyond
this, but if a Partitioner is specified, we would not change the number of
partitions.

Some initial questions we had:
 * Does spark already do this?
 * Is there interest in supporting this functionality?
 * Are there any potential issues that we should be aware of?
 * What changes would need to be made for such a project?

Thanks,
Muhammed

Reply via email to