Hi Eric,

You can run a job from another one, using the REST API
This is the only way we have found to launch a batch job from a streaming
job

------------------

Bastien DINE
Data Architect / Software Engineer / Sysadmin
bastiendine.io


Le ven. 23 nov. 2018 à 11:52, Piotr Nowojski <pi...@data-artisans.com> a
écrit :

> Hi,
>
> I’m not sure if I understand your problem and your context, but spawning a
> batch job every 45 seconds doesn’t sound as a that bad idea (as long as the
> job is short).
>
> Another idea would be to incorporate this batch job inside your streaming
> job, for example by reading from Cassandra using an AsyncIO operator:
>
> https://ci.apache.org/projects/flink/flink-docs-release-1.6/dev/stream/operators/asyncio.html
>
> Quick google search revealed for example this:
>
>
> https://stackoverflow.com/questions/43067681/read-data-from-cassandra-for-processing-in-flink
>
> Piotrek
>
> > On 23 Nov 2018, at 10:33, eric hoffmann <sigfrid.hoffm...@gmail.com>
> wrote:
> >
> > Hi
> > Is it possible to call batch job on a streaming context?
> > what i want to do is:
> > for a given input event, fetch cassandra elements based on event data,
> apply transformation on them and apply a ranking when all elements fetched
> by cassandra are processed.
> > If i do this in batch mode i would have to submit a job on each events
> and i can have an event every 45 seconds.
> > Is there any alternative? can i start a batch job that will receive some
> external request, process it and wait for another request?
> > thx
> > Eric
>
>

Reply via email to