One possible workaround could be to orchestrate launch/stopping of Streaming jobs on demand as long as the number of jobs/streams stay within the boundaries of the resources (cores) you've available. e.g. if you're using Mesos, Marathon offers a REST interface to manage job lifecycle. You will still need to solve the dynamic configuration through some alternative channel.
On Wed, Jan 21, 2015 at 11:30 AM, Tamas Jambor <jambo...@gmail.com> wrote: > thanks for the replies. > > is this something we can get around? Tried to hack into the code without > much success. > > On Wed, Jan 21, 2015 at 3:15 AM, Shao, Saisai <saisai.s...@intel.com> > wrote: > >> Hi, >> >> I don't think current Spark Streaming support this feature, all the >> DStream lineage is fixed after the context is started. >> >> Also stopping a stream is not supported, instead currently we need to >> stop the whole streaming context to meet what you want. >> >> Thanks >> Saisai >> >> -----Original Message----- >> From: jamborta [mailto:jambo...@gmail.com] >> Sent: Wednesday, January 21, 2015 3:09 AM >> To: user@spark.apache.org >> Subject: dynamically change receiver for a spark stream >> >> Hi all, >> >> we have been trying to setup a stream using a custom receiver that would >> pick up data from sql databases. we'd like to keep that stream context >> running and dynamically change the streams on demand, adding and removing >> streams based on demand. alternativel, if a stream is fixed, is it possible >> to stop a stream, change to config and start again? >> >> thanks, >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/dynamically-change-receiver-for-a-spark-stream-tp21268.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional >> commands, e-mail: user-h...@spark.apache.org >> >> >