Re: Access Sliding window

2017-08-07 Thread Raj Kumar
Hi Fabian, Can you please answer my last set of questions I have posted on the Forum. Thanks. On Friday, August 4, 2017, Fabian Hueske-2 [via Apache Flink User Mailing List archive.] wrote: > TimeWindow.getStart() or TimeWindow.getEnd() > > -> https://ci.apache.org/projects/flink/flink-docs- >

Re: JMX stats reporter with all task manager/job manager stats aggregated?

2017-08-07 Thread Chesnay Schepler
Hello, there is no central place where JMX metrics are aggregated. You can configure a port range for the reporter to prevent port conflicts on the same machine. metrics.reporter.jmx.port:8789-8790 You can find out which port was used by checking the logs. Regards, Chesnay On 05.08.2017 03

Re: [ANNOUNCE] Apache Flink 1.3.2 released

2017-08-07 Thread Till Rohrmann
Thanks to the community and Aljoscha for the hard work to complete Flink 1.3.2. Cheers, Till On Sat, Aug 5, 2017 at 9:12 AM, Aljoscha Krettek wrote: > The Apache Flink community is pleased to announce the release of Apache > Flink 1.3.2. > > Apache Flink® is an open-source stream processing fra

Re: [ANNOUNCE] Apache Flink 1.3.2 released

2017-08-07 Thread Fabian Hueske
Thanks Aljoscha and everybody who contributed with bug reports and fixes! Best, Fabian 2017-08-07 11:07 GMT+02:00 Till Rohrmann : > Thanks to the community and Aljoscha for the hard work to complete Flink > 1.3.2. > > Cheers, > Till > > On Sat, Aug 5, 2017 at 9:12 AM, Aljoscha Krettek > wrote:

Re: Test example and lambdas

2017-08-07 Thread Timo Walther
Hi, Row is very special data type, because Flink cannot extract the field types automatically based on Java generics. By default it is serialized by Kryo, you need to specify the field types using Typles.ROW(Types.STRING, ...) and pass this information in your `.returns()` methods instead of

Flink REST API async?

2017-08-07 Thread Francisco Gonzalez Barea
Hi there! We are doing some POCs submitting jobs remotely to Flink. We tried with Flink CLI and now we´re testing the Rest API. So the point is that when we try to execute a set of requests in an async way (using CompletableFutures) only a couple of them run successfully. For the rest we get t

Re: [EXTERNAL] Re: Help required - "BucketingSink" usage to write HDFS Files

2017-08-07 Thread Raja . Aravapalli
Thanks very much for the pointers Vinay. That helps ☺ -Raja. From: vinay patil Date: Monday, August 7, 2017 at 1:56 AM To: "user@flink.apache.org" Subject: Re: [EXTERNAL] Re: Help required - "BucketingSink" usage to write HDFS Files Hi Raja, That is why they are in the pending state. You ca

schema to just read as "byte[] array" from kafka

2017-08-07 Thread Raja . Aravapalli
Hi I am using SimpleStringSchema to deserialize a message read from kafka, but need some help to know if there is any schema available I can use rather than “SimpleStringSchema()” and instead just get “byte[]” without any deserialization happening! Below is code I am currently using, but inst

Re: Flink REST API async?

2017-08-07 Thread Eron Wright
When you submit a program via the REST API, the main method executes inside the JobManager process.Unfortunately a static variable is used to establish the execution environment that the program obtains from `ExecutionEnvironment.getExecutionEnvironment()`. From the stack trace it appears that

Re: WaterMark & Eventwindow not fired correctly

2017-08-07 Thread aitozi
Hi, my flink version is 1.2 i am work on this problem these days. Below is my found. when i use "assignTimestampsAndWatermark" with same parallelism as 240 as the before operator, the before operator has two input(it is a "connected" Co-FlatMap operator with parallelism 240), it runs into that

Re: schema to just read as "byte[] array" from kafka

2017-08-07 Thread Chao Wang
Hi Raja, I just happened to work on the similar thing, and here is how to do it in general, I think (In my case, I did a bit more, to deserialize a tuple of ) : FlinkKafkaConsumer010 consumer = new FlinkKafkaConsumer010<>("topic_name", new MyDe(), properties); and for MyDe the schema: pub

Re: schema to just read as "byte[] array" from kafka

2017-08-07 Thread Chao Wang
A quick update, in class MyDe: public static class MyDe extends AbstractDeserializationSchema { @Override public byte[] deserialize(byte[] arg0) { // Perform deserialization here, if needed; // otherwise, probably we can simply return arg0 as raw byte[] return arg0; } } Chao

Re: [EXTERNAL] Re: schema to just read as "byte[] array" from kafka

2017-08-07 Thread Raja . Aravapalli
Thank you very much Chao. That helps me. Regards, Raja. From: Chao Wang Date: Monday, August 7, 2017 at 12:28 PM To: Raja Aravapalli Cc: "user@flink.apache.org" Subject: [EXTERNAL] Re: schema to just read as "byte[] array" from kafka A quick update, in class MyDe: public static class MyDe

load + update global state

2017-08-07 Thread Peter Ertl
Hi folks, I am coding a streaming task that processes http requests from our web site and enriches these with additional information. It contains session ids from historic requests and the related emails that were used within these session in the past. lookup - hashtable: session_id:

Experiencing long latency while using sockets

2017-08-07 Thread Chao Wang
Hi, I have been trying to benchmark the end-to-end latency of a Flink 1.3.1 application, but got confused regarding the amount of time spent in Flink. In my setting, data source and data sink dwell in separated machines, like the following topology: Machine 1

Re: Experiencing long latency while using sockets

2017-08-07 Thread Chao Wang
Following the original post, I've tried stripping down my Flink app to only the following, and then it still exhibits long latencies: after the second source socket write, it took 90+ milliseconds from data source to the socket-front in Flink. I would like to ask for pointers about how to inves

Flink streaming Parallelism

2017-08-07 Thread Jakes John
I am coming from Apache Storm world. I am planning to switch from storm to flink. I was reading Flink documentation but, I couldn't find some requirements in Flink which was present in Storm. I need to have a streaming pipeline Kafka->flink-> ElasticSearch. In storm, I have seen that I

Re: Flink streaming Parallelism

2017-08-07 Thread Tzu-Li (Gordon) Tai
Hi, The equivalent would be setting a parallelism on your sink operator. e.g. stream.addSink(…).setParallelism(…). By default the parallelism of all operators in the pipeline will be whatever parallelism was set for the whole job, unless parallelism is explicitly set for a specific operator. Fo

Re: load + update global state

2017-08-07 Thread Tzu-Li (Gordon) Tai
Hi Peter! One thing I’d like to understand first after reading about your use case: Why exactly do you need the lookup table to be globally accessible? From what I understand, you are using this lookup table for stream event enriching, so whatever processing you need to perform downstream on thi