spark.streaming.concurrentJobs is a driver side internal configuration, this means that how many streaming jobs can be submitted concurrently in one batch. Usually this should not be configured by user, unless you're familiar with Spark Streaming internals, and know the implication of this configuration.
thomas lavocat <thomas.lavo...@univ-grenoble-alpes.fr> 于2018年6月5日周二 下午4:20写道: > Hi everyone, > > I'm wondering if the property spark.streaming.concurrentJobs should > reflects the total number of possible concurrent task on the cluster, or > the a local number of concurrent tasks on one compute node. > > Thanks for your help. > > Thomas > > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >