Hi Fabian,
Can you please answer my last set of questions I have posted on the Forum.
Thanks.
On Friday, August 4, 2017, Fabian Hueske-2 [via Apache Flink User Mailing
List archive.] wrote:
> TimeWindow.getStart() or TimeWindow.getEnd()
>
> -> https://ci.apache.org/projects/flink/flink-docs-
>
Hello,
there is no central place where JMX metrics are aggregated.
You can configure a port range for the reporter to prevent port
conflicts on the same machine.
metrics.reporter.jmx.port:8789-8790
You can find out which port was used by checking the logs.
Regards,
Chesnay
On 05.08.2017 03
Thanks to the community and Aljoscha for the hard work to complete Flink
1.3.2.
Cheers,
Till
On Sat, Aug 5, 2017 at 9:12 AM, Aljoscha Krettek
wrote:
> The Apache Flink community is pleased to announce the release of Apache
> Flink 1.3.2.
>
> Apache Flink® is an open-source stream processing fra
Thanks Aljoscha and everybody who contributed with bug reports and fixes!
Best, Fabian
2017-08-07 11:07 GMT+02:00 Till Rohrmann :
> Thanks to the community and Aljoscha for the hard work to complete Flink
> 1.3.2.
>
> Cheers,
> Till
>
> On Sat, Aug 5, 2017 at 9:12 AM, Aljoscha Krettek
> wrote:
Hi,
Row is very special data type, because Flink cannot extract the field
types automatically based on Java generics. By default it is serialized
by Kryo, you need to specify the field types using
Typles.ROW(Types.STRING, ...) and pass this information in your
`.returns()` methods instead of
Hi there!
We are doing some POCs submitting jobs remotely to Flink. We tried with Flink
CLI and now we´re testing the Rest API.
So the point is that when we try to execute a set of requests in an async way
(using CompletableFutures) only a couple of them run successfully. For the rest
we get t
Thanks very much for the pointers Vinay. That helps ☺
-Raja.
From: vinay patil
Date: Monday, August 7, 2017 at 1:56 AM
To: "user@flink.apache.org"
Subject: Re: [EXTERNAL] Re: Help required - "BucketingSink" usage to write HDFS
Files
Hi Raja,
That is why they are in the pending state. You ca
Hi
I am using SimpleStringSchema to deserialize a message read from kafka, but
need some help to know if there is any schema available I can use rather than
“SimpleStringSchema()” and instead just get “byte[]” without any
deserialization happening!
Below is code I am currently using, but inst
When you submit a program via the REST API, the main method executes inside
the JobManager process.Unfortunately a static variable is used to
establish the execution environment that the program obtains from
`ExecutionEnvironment.getExecutionEnvironment()`. From the stack trace it
appears that
Hi,
my flink version is 1.2
i am work on this problem these days. Below is my found.
when i use "assignTimestampsAndWatermark" with same parallelism as 240 as
the before operator, the before operator has two input(it is a "connected"
Co-FlatMap operator with parallelism 240), it runs into that
Hi Raja,
I just happened to work on the similar thing, and here is how to do it
in general, I think (In my case, I did a bit more, to deserialize a
tuple of ) :
FlinkKafkaConsumer010 consumer = new
FlinkKafkaConsumer010<>("topic_name", new MyDe(), properties);
and for MyDe the schema:
pub
A quick update, in class MyDe:
public static class MyDe extends AbstractDeserializationSchema {
@Override
public byte[] deserialize(byte[] arg0) {
// Perform deserialization here, if needed;
// otherwise, probably we can simply return arg0 as raw byte[]
return arg0;
}
}
Chao
Thank you very much Chao. That helps me.
Regards,
Raja.
From: Chao Wang
Date: Monday, August 7, 2017 at 12:28 PM
To: Raja Aravapalli
Cc: "user@flink.apache.org"
Subject: [EXTERNAL] Re: schema to just read as "byte[] array" from kafka
A quick update, in class MyDe:
public static class MyDe
Hi folks,
I am coding a streaming task that processes http requests from our web site and
enriches these with additional information.
It contains session ids from historic requests and the related emails that were
used within these session in the past.
lookup - hashtable: session_id:
Hi,
I have been trying to benchmark the end-to-end latency of a Flink 1.3.1
application, but got confused regarding the amount of time spent in
Flink. In my setting, data source and data sink dwell in separated
machines, like the following topology:
Machine 1
Following the original post, I've tried stripping down my Flink app to
only the following, and then it still exhibits long latencies: after the
second source socket write, it took 90+ milliseconds from data source to
the socket-front in Flink. I would like to ask for pointers about how to
inves
I am coming from Apache Storm world. I am planning to switch from
storm to flink. I was reading Flink documentation but, I couldn't find some
requirements in Flink which was present in Storm.
I need to have a streaming pipeline Kafka->flink-> ElasticSearch. In
storm, I have seen that I
Hi,
The equivalent would be setting a parallelism on your sink operator. e.g.
stream.addSink(…).setParallelism(…).
By default the parallelism of all operators in the pipeline will be whatever
parallelism was set for the whole job, unless parallelism is explicitly set for
a specific operator. Fo
Hi Peter!
One thing I’d like to understand first after reading about your use case:
Why exactly do you need the lookup table to be globally accessible? From what I
understand, you are using this lookup table for stream event enriching, so
whatever processing you need to perform downstream on thi
19 matches
Mail list logo