Hi Vishal,

my approach would be a single Kubernetes service, which is backed by all
Taskmanagers of the job. The Taskmanagers will proxy the request for a
specific key to the correct Taskmanager. Yes, the Taskmanagers will cache
the location of the key groups.

In addition to this Kubernetes service, you can of course have a
Jetty/Jersey REST based server that sends queries to this service.

Please le me know if this works for you.

Hope this helps and cheers,

Konstantin


On Thu, Mar 28, 2019 at 12:37 AM Vishal Santoshi <vishal.santo...@gmail.com>
wrote:

> I think I got a handle on this. For those who might want to do this
>
>
> Here are the steps ( I could share the  Jetty/Jersey REST code too is
> required )
>
>
> *1.* Create a side car container on each pod that has a TM. I wrote a
> simple Jetty/Jersey REST based server that  execute queries against the
> local TM query server.
>
>   .
>
>   .
>
>   - name: queryable-state
>
>         image: _IMAGE_
>
>         args: ["queryable-state"]
>
>         env:
>
>           - name: POD_IP
>
>             valueFrom:
>
>               fieldRef:
>
>                 fieldPath: status.podIP
>
>         ports:
>
>           - containerPort: 9999
>
>             name: qstate-client
>
>         resources:
>
>           requests:
>
>             cpu: "0.25"
>
>             memory: "256Mi"
>
>
>    Note that POD_IP is the ip used by the REST based server to start the
> QueryableStateClient and the port is the default port of the TM query
> server ( 9069 I think ) of the colocated TM container.
>
>
>
> *2.* Expose the port ( in this case 9999 ) at the k8s service layer.
>
>
>
> And that did it.
>
>
>
>
>
>
>
>
> I though am worried about a couple of things
>
>
> *1*.
>
>  The  TM query server will ask JM for the key group and hence the TM a
> key belongs to for every request and then coordinate the coummunication
> between the client and that TM. Does flink do any optimzation, as in cache
> the key ranges and thus the affinity to a TM to reduce JM stress. I would
> imagine that being some well known distribution function on some well known
> hash algorithm, an incident key could be pinned to a TM without visiting
> the JM more then once.
>
>
> *2. *
>
> We do have use cases where we would want to iterate over all the keys in a
> key group ( and by extension on a TM ) for a job. Is that a possibility ?
>
>
> *3. *
>
> The overhead of having as many client containers as TMs.
>
>
>
> Any advise/ideas on the 3 worry points ?
>
>
>
> Regards
>
> On Mon, Mar 25, 2019 at 8:57 PM Vishal Santoshi <vishal.santo...@gmail.com>
> wrote:
>
>> I  have 2 options
>>
>> 1. A Rest Based,  in my case a Jetty/REST based QueryableStateClient  in
>> a side car container colocated on JM  ( Though it could on all TMs but that
>> looks to an overkill )
>>
>> 2.A Rest Based,  in my case a Jetty/REST based QueryableStateClient  in a
>> side car container colocated on TMs.  The Query Proxies are on the TMs, so
>> in essence the communication would be within containers of the POD and I
>> could load balance ( have ot test  )
>>
>> The second alternative seems doable, but looks an overkill  but am not
>> sure how to establish a TM on the standalone QueryableStateClient, given
>> that TM's pod IP is not known till the pod is launched.
>>
>> Has anyone had a successful QueryableState setup for flink  on k8s?
>>
>> Regards,
>>
>

-- 

Konstantin Knauf | Solutions Architect

+49 160 91394525

<https://www.ververica.com/>

Follow us @VervericaData

--

Join Flink Forward <https://flink-forward.org/> - The Apache Flink
Conference

Stream Processing | Event Driven | Real Time

--

Data Artisans GmbH | Invalidenstrasse 115, 10115 Berlin, Germany

--
Data Artisans GmbH
Registered at Amtsgericht Charlottenburg: HRB 158244 B
Managing Directors: Dr. Kostas Tzoumas, Dr. Stephan Ewen

Reply via email to