Hi Kostas,

Yes. Exactly. Thanks a lot for this one.

That's really what we need !


Cheers

On Sun, Nov 15, 2015 at 8:53 PM, Kostas Tzoumas <ktzou...@apache.org> wrote:

> Hi Wally,
>
> This version adds support for specifying and switching between time
> semantics - processing time, ingestion time, or event time.
>
> When working with event time, you can specify watermarks to track the
> progress of event time. So, even if events arrive out of order, windows
> will be specified on the event time (not arrival time), and the computation
> will be triggered on watermark arrival.
>
> You can see the API reference and an example here:
> https://ci.apache.org/projects/flink/flink-docs-release-0.10/apis/streaming_guide.html#working-with-time
>
> Is this what you are looking for?
>
> Kostas
>
>
> On Sat, Nov 14, 2015 at 1:54 AM, Welly Tambunan <if05...@gmail.com> wrote:
>
>> Hi Robert,
>>
>> Is this version has already handle the stream perfection or out of order
>> event ?
>>
>> Any resource on how this work and the API reference ?
>>
>>
>> Cheers
>>
>> On Fri, Nov 13, 2015 at 4:00 PM, Welly Tambunan <if05...@gmail.com>
>> wrote:
>>
>>> Awesome !
>>>
>>> This is really the best weekend gift ever. :)
>>>
>>> Cheers
>>>
>>> On Fri, Nov 13, 2015 at 3:54 PM, Robert Metzger <rmetz...@apache.org>
>>> wrote:
>>>
>>>> Hi Welly,
>>>> Flink 0.10.0 is out, its just not announced yet.
>>>> Its available on maven central and the global mirrors are currently
>>>> syncing it. This mirror for example has the update already:
>>>> http://apache.mirror.digionline.de/flink/flink-0.10.0/
>>>>
>>>> On Fri, Nov 13, 2015 at 9:50 AM, Welly Tambunan <if05...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Aljoscha,
>>>>>
>>>>> Thanks for this one. Looking forward for 0.10 release version.
>>>>>
>>>>> Cheers
>>>>>
>>>>> On Thu, Nov 12, 2015 at 5:34 PM, Aljoscha Krettek <aljos...@apache.org
>>>>> > wrote:
>>>>>
>>>>>> Hi,
>>>>>> I don’t know yet when the operator state will be transitioned to
>>>>>> managed memory but it could happen for 1.0 (which will come after 0.10).
>>>>>> The good thing is that the interfaces won’t change, so state can be used 
>>>>>> as
>>>>>> it is now.
>>>>>>
>>>>>> For 0.10, the release vote is winding down right now, so you can
>>>>>> expect the release to happen today or tomorrow. I think the streaming is
>>>>>> production ready now, we expect to mostly to hardening and some
>>>>>> infrastructure changes (for example annotations that specify API 
>>>>>> stability)
>>>>>> for the 1.0 release.
>>>>>>
>>>>>> Let us know if you need more information.
>>>>>>
>>>>>> Cheers,
>>>>>> Aljoscha
>>>>>> > On 12 Nov 2015, at 02:42, Welly Tambunan <if05...@gmail.com> wrote:
>>>>>> >
>>>>>> > Hi Stephan,
>>>>>> >
>>>>>> > >Storing the model in OperatorState is a good idea, if you can. On
>>>>>> the roadmap is to migrate the operator state to managed memory as well, 
>>>>>> so
>>>>>> that should take care of the GC issues.
>>>>>> > Is this using off the heap memory ? Which version we expect this
>>>>>> one to be available ?
>>>>>> >
>>>>>> > Another question is when will the release version of 0.10 will be
>>>>>> out ? We would love to upgrade to that one when it's available. That
>>>>>> version will be a production ready streaming right ?
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > On Wed, Nov 11, 2015 at 4:49 PM, Stephan Ewen <se...@apache.org>
>>>>>> wrote:
>>>>>> > Hi!
>>>>>> >
>>>>>> > In general, if you can keep state in Flink, you get better
>>>>>> throughput/latency/consistency and have one less system to worry about
>>>>>> (external k/v store). State outside means that the Flink processes can be
>>>>>> slimmer and need fewer resources and as such recover a bit faster. There
>>>>>> are use cases for that as well.
>>>>>> >
>>>>>> > Storing the model in OperatorState is a good idea, if you can. On
>>>>>> the roadmap is to migrate the operator state to managed memory as well, 
>>>>>> so
>>>>>> that should take care of the GC issues.
>>>>>> >
>>>>>> > We are just adding functionality to make the Key/Value operator
>>>>>> state usable in CoMap/CoFlatMap as well (currently it only works in 
>>>>>> windows
>>>>>> and in Map/FlatMap/Filter functions over the KeyedStream).
>>>>>> > Until the, you should be able to use a simple Java HashMap and use
>>>>>> the "Checkpointed" interface to get it persistent.
>>>>>> >
>>>>>> > Greetings,
>>>>>> > Stephan
>>>>>> >
>>>>>> >
>>>>>> > On Sun, Nov 8, 2015 at 10:11 AM, Welly Tambunan <if05...@gmail.com>
>>>>>> wrote:
>>>>>> > Thanks for the answer.
>>>>>> >
>>>>>> > Currently the approach that i'm using right now is creating a
>>>>>> base/marker interface to stream different type of message to the same
>>>>>> operator. Not sure about the performance hit about this compare to the
>>>>>> CoFlatMap function.
>>>>>> >
>>>>>> > Basically this one is providing query cache, so i'm thinking
>>>>>> instead of using in memory cache like redis, ignite etc, i can just use
>>>>>> operator state for this one.
>>>>>> >
>>>>>> > I just want to gauge do i need to use memory cache or operator
>>>>>> state would be just fine.
>>>>>> >
>>>>>> > However i'm concern about the Gen 2 Garbage Collection for caching
>>>>>> our own state without using operator state. Is there any clarification on
>>>>>> that one ?
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > On Sat, Nov 7, 2015 at 12:38 AM, Anwar Rizal <anriza...@gmail.com>
>>>>>> wrote:
>>>>>> >
>>>>>> > Let me understand your case better here. You have a stream of model
>>>>>> and stream of data. To process the data, you will need a way to access 
>>>>>> your
>>>>>> model from the subsequent stream operations (map, filter, flatmap, ..).
>>>>>> > I'm not sure in which case Operator State is a good choice, but I
>>>>>> think you can also live without.
>>>>>> >
>>>>>> > val modelStream = .... // get the model stream
>>>>>> > val dataStream   =
>>>>>> >
>>>>>> > modelStream.broadcast.connect(dataStream). coFlatMap(  ) Then you
>>>>>> can keep the latest model in a CoFlatMapRichFunction, not necessarily as
>>>>>> Operator State, although maybe OperatorState is a good choice too.
>>>>>> >
>>>>>> > Does it make sense to you ?
>>>>>> >
>>>>>> > Anwar
>>>>>> >
>>>>>> > On Fri, Nov 6, 2015 at 10:21 AM, Welly Tambunan <if05...@gmail.com>
>>>>>> wrote:
>>>>>> > Hi All,
>>>>>> >
>>>>>> > We have a high density data that required a downsample. However
>>>>>> this downsample model is very flexible based on the client device and 
>>>>>> user
>>>>>> interaction. So it will be wasteful to precompute and store to db.
>>>>>> >
>>>>>> > So we want to use Apache Flink to do downsampling and cache the
>>>>>> result for subsequent query.
>>>>>> >
>>>>>> > We are considering using Flink Operator state for that one.
>>>>>> >
>>>>>> > Is that the right approach to use that for memory cache ? Or if
>>>>>> that preferable using memory cache like redis etc.
>>>>>> >
>>>>>> > Any comments will be appreciated.
>>>>>> >
>>>>>> >
>>>>>> > Cheers
>>>>>> > --
>>>>>> > Welly Tambunan
>>>>>> > Triplelands
>>>>>> >
>>>>>> > http://weltam.wordpress.com
>>>>>> > http://www.triplelands.com
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > --
>>>>>> > Welly Tambunan
>>>>>> > Triplelands
>>>>>> >
>>>>>> > http://weltam.wordpress.com
>>>>>> > http://www.triplelands.com
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > --
>>>>>> > Welly Tambunan
>>>>>> > Triplelands
>>>>>> >
>>>>>> > http://weltam.wordpress.com
>>>>>> > http://www.triplelands.com
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Welly Tambunan
>>>>> Triplelands
>>>>>
>>>>> http://weltam.wordpress.com
>>>>> http://www.triplelands.com <http://www.triplelands.com/blog/>
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Welly Tambunan
>>> Triplelands
>>>
>>> http://weltam.wordpress.com
>>> http://www.triplelands.com <http://www.triplelands.com/blog/>
>>>
>>
>>
>>
>> --
>> Welly Tambunan
>> Triplelands
>>
>> http://weltam.wordpress.com
>> http://www.triplelands.com <http://www.triplelands.com/blog/>
>>
>
>


-- 
Welly Tambunan
Triplelands

http://weltam.wordpress.com
http://www.triplelands.com <http://www.triplelands.com/blog/>

Reply via email to