No, that applies to the streaming DataFrame API too. No jobs can't communicate with each other.
On Tue, Aug 24, 2021 at 9:51 PM Artemis User <arte...@dtechspace.com> wrote: > Thanks Daniel. I guess you were suggesting using DStream/RDD. Would it > be possible to use structured streaming/DataFrames for multi-source > streaming? In addition, we really need each stream data ingestion to be > asynchronous or non-blocking... thanks! > > On 8/24/21 9:27 PM, daniel williams wrote: > > Yeah. Build up the streams as a collection and map that query to the > start() invocation and map those results to awaitTermination() or whatever > other blocking mechanism you’d like to use. > > On Tue, Aug 24, 2021 at 4:37 PM Artemis User <arte...@dtechspace.com> > wrote: > >> Is there a way to run multiple streams in a single Spark job using >> Structured Streaming? If not, is there an easy way to perform inter-job >> communications (e.g. referencing a dataframe among concurrent jobs) in >> Spark? Thanks a lot in advance! >> >> -- ND >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >> -- > -dan > > >