Hi Madhukara,

What I understood from the code is that when ever runBatch return they
trigger constructBatch so whatever is processing time for a batch will be
ur batch time if u dnt specify a trigger.

one flaw which i think in this is if your processing time keeps increasing
with amount of data , then this batch interval keeps on increasing, they
must put some boundary or some logic to block to prevent such case.

here is one jira which i found related to this:-
https://github.com/apache/spark/pull/12725


On Fri, May 6, 2016 at 2:50 PM, Deepak Sharma <deepakmc...@gmail.com> wrote:

> With Structured Streaming ,Spark would provide apis over spark sql engine.
> Its like once you have the structured stream and dataframe created out of
> this , you can do ad-hoc querying on the DF , which means you are actually
> querying the stram without having to store or transform.
> I have not used it yet but seems it will be like start streaming data from
> source  as son as you define it.
>
> Thanks
> Deepak
>
>
> On Fri, May 6, 2016 at 1:37 PM, madhu phatak <phatak....@gmail.com> wrote:
>
>> Hi,
>> As I was playing with new structured streaming API, I noticed that spark
>> starts processing as and when the data appears. It's no more seems like
>> micro batch processing. Is spark structured streaming will be an event
>> based processing?
>>
>> --
>> Regards,
>> Madhukara Phatak
>> http://datamantra.io/
>>
>
>
>
> --
> Thanks
> Deepak
> www.bigdatabig.com
> www.keosha.net
>



-- 

Thanks & Regards

Sachin Aggarwal
7760502772

Reply via email to