Thanks Sean. Excuse my ignorant, but I just can't figure out how to
create a collection across multiple streams using multiple stream
readers. Could you provide some examples or additional references? Thanks!
On 8/24/21 11:01 PM, Sean Owen wrote:
No, that applies to the streaming DataFrame API too.
No jobs can't communicate with each other.
On Tue, Aug 24, 2021 at 9:51 PM Artemis User <arte...@dtechspace.com
<mailto:arte...@dtechspace.com>> wrote:
Thanks Daniel. I guess you were suggesting using DStream/RDD.
Would it be possible to use structured streaming/DataFrames for
multi-source streaming? In addition, we really need each stream
data ingestion to be asynchronous or non-blocking... thanks!
On 8/24/21 9:27 PM, daniel williams wrote:
Yeah. Build up the streams as a collection and map that query to
the start() invocation and map those results to
awaitTermination() or whatever other blocking mechanism you’d
like to use.
On Tue, Aug 24, 2021 at 4:37 PM Artemis User
<arte...@dtechspace.com <mailto:arte...@dtechspace.com>> wrote:
Is there a way to run multiple streams in a single Spark job
using
Structured Streaming? If not, is there an easy way to
perform inter-job
communications (e.g. referencing a dataframe among concurrent
jobs) in
Spark? Thanks a lot in advance!
-- ND
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org
<mailto:user-unsubscr...@spark.apache.org>
--
-dan