Re: About KafkaConsumerBase

2017-08-01 Thread Tzu-Li (Gordon) Tai
Hi, it maintain itself a individual instance of  FlinkKafkaConsumerBase, and it contains a individual pendingOffsetsToCommit  , am right ?  That is correct! The FlinkKafkaConsumerBase is code executed for each parallel subtask instance, and therefore have their own pendingOffsetsToCommit which w

Re: About KafkaConsumerBase

2017-08-01 Thread aitozi
Hi,Piotr Nowojski i think you are right, but i think it is executed in parallel, but in each parallel , it maintain itself a individual instance of FlinkKafkaConsumerBase, and it contains a individual pendingOffsetsToCommit , am right ? thanks, aitozi -- View this message in context:

Eventime window

2017-08-01 Thread Govindarajan Srinivasaraghavan
Hi, I have few questions regarding event time windowing. My scenario is devices from various timezones will send messages with timestamp and I need to create a window per device for 10 seconds. The messages will mostly arrive in order. Here is my sample code to perform windowing and aggregating t

Re: Using FileInputFormat - org.apache.flink.streaming.api.functions.source.TimestampedFileInputSplit

2017-08-01 Thread Mohit Anchlia
This was user induced problem - me. I wasn't calling streamenv.execute() :( On Tue, Aug 1, 2017 at 1:29 PM, Mohit Anchlia wrote: > This doesn't work even with TextInputFormat. Not sure what's wrong. > > On Tue, Aug 1, 2017 at 9:53 AM, Mohit Anchlia > wrote: > >> I don't see the print output. >>

Re: Using FileInputFormat - org.apache.flink.streaming.api.functions.source.TimestampedFileInputSplit

2017-08-01 Thread Mohit Anchlia
This doesn't work even with TextInputFormat. Not sure what's wrong. On Tue, Aug 1, 2017 at 9:53 AM, Mohit Anchlia wrote: > I don't see the print output. > > On Tue, Aug 1, 2017 at 2:08 AM, Fabian Hueske wrote: > >> Hi Mohit, >> >> these are just INFO log statements that do not necessarily indic

S3 Write Execption

2017-08-01 Thread Aneesha Kaushal
Hello, I am using flink 1.2 and writing records to S3 using rolling sink. I am encountering this S3 write error quite frequently : TimerException{com.amazonaws.services.s3.model.AmazonS3Exception: Status Code: 404, AWS Service: Amazon S3, AWS Request ID: B573887B1850BF28, AWS Error Code: nu

Re: Odd flink behaviour

2017-08-01 Thread Mohit Anchlia
Thanks that worked. However, what I don't understand is wouldn't the open call that I am inheriting have this logic already inbuilt? I am inheriting FileInputFormat. On Tue, Aug 1, 2017 at 1:42 AM, Fabian Hueske wrote: > An InputFormat processes multiple InputSplits. open() is called for each >

Proper way to establish bucket counts

2017-08-01 Thread Robert Rapplean
I want a count of events that are put into a bucketing sink, but can't find a ready-made way of doing that. Is there an easier way than to implement a counter for each bucket via the metrics? If metrics counters is the easy way, what do I do to make sure that I don't have a memory leak from expired

Re: Using FileInputFormat - org.apache.flink.streaming.api.functions.source.TimestampedFileInputSplit

2017-08-01 Thread Mohit Anchlia
I don't see the print output. On Tue, Aug 1, 2017 at 2:08 AM, Fabian Hueske wrote: > Hi Mohit, > > these are just INFO log statements that do not necessarily indicate a > problem. > Is the program working otherwise or do you observe other problems? > > Best, Fabian > > 2017-08-01 0:32 GMT+02:00

Re: Azkaban Job Type Plugin for Flink

2017-08-01 Thread Aljoscha Krettek
Hi Yann, I'm not aware of anyone that has started work on this. I'm quite interested in how this turns out for you. Best, Aljoscha > On 20. Jul 2017, at 14:49, Yann Pauly wrote: > > Hi all, > > We want to integrate our Flink instances with our Azkaban scheduler. > For that we will have to cr

Re: Is that possible for flink to dynamically read and change configuration?

2017-08-01 Thread Aljoscha Krettek
Just some clarification: Flink state is never shared between different parallel operator instances. If you want to make those changes available to all parallel instances of the operation you have to broadcast the control stream, i.e. control.broadcast(). Best, Aljoscha > On 24. Jul 2017, at 17

Re: Flink CLI cannot submit job to Flink on Mesos

2017-08-01 Thread Stephan Ewen
Cool, good to hear! It is one of those "it a feature, not a bug" situations ;-) Flink's HA mode supports multiple masters, so the CLI needs to have a way to find which master is "leader" (active, versus the passive masters on standby). That discovery goes through ZooKeeper as well (which is the g

Re: Split Streams not working

2017-08-01 Thread Aljoscha Krettek
Hi, In your original program, the problem is that there is both an ERROR and SUCCESS event in your List. Thus you add both "success" and "error" to the list of split outputs. To discern between those different types you first have to flatten that DataStream> into a DataStream using a flatMap().

Re: Class not found when deserializing

2017-08-01 Thread Aljoscha Krettek
Hi, I think the problem here is that the SerializationUtils don't use the correct ClassLoader when deserialising. When running a Flink job on a cluster the user code (from the user-code jar) is only available in the user-code ClassLoader. If you can pass a ClassLoader to your deserialize method

Re: Flink QueryableState with Sliding Window on RocksDB

2017-08-01 Thread Biplob Biswas
Hi Fabian, I am not really sure using CoProcessFunction would be useful for my use case. My use case, in short, can be explained as follows: 1) create 2 different local state store, where both have 1-N relationship. For eg. 1 -> [A,B,C] and A -> [1,2,3] 2) Based on the key A, get list of element

Re: SQL API with Table API

2017-08-01 Thread nragon
"No, those are two different queries. " This is enough. The second part does not applies since i'm calculating EventTime from table source. Thanks -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/SQL-API-with-Table-API-tp14599p14605.html Sen

Re: SQL API with Table API

2017-08-01 Thread Fabian Hueske
No, those are two different queries. The second query would also not work because the revenue table does not have the EventTime attribute. Best, Fabian 2017-08-01 13:03 GMT+02:00 nragon : > Hi, > > Can i expect the output from this: > > Table revenue = tableEnv.sql( > "SELECT TUMBLE_START(E

Re: About KafkaConsumerBase

2017-08-01 Thread Piotr Nowojski
Hi, pendingOffsetsToCommit is a private field which is not accessed from outside of the FlinkKafkaConsumerBase class. It is only used in state manipulation methods, which are not executed in parallel. Thanks, Piotrek > On Aug 1, 2017, at 1:20 PM, aitozi wrote: > > Hello: > >i am new to

Re: Invalid path exception

2017-08-01 Thread Chesnay Schepler
Let's move the discussions to FLINK-7330. On 01.08.2017 13:15, Chesnay Schepler wrote: One problem i know of is that windows paths with a scheme are not detected as windows paths, as documented in FLINK-6889. They generally still work though (/maybe /by chance). I just verified that calling Fi

About KafkaConsumerBase

2017-08-01 Thread aitozi
Hello: i am new to Flink, ijust read the source code . i am doubt that , why in FlinkKafkaConsumerBase.java (version1.2), like method : notifyCheckpointComplete may change the pendingOffsetsToCommit in parallel , why dont need to be surrouned with "synchronized" thanks -- View this me

Re: Invalid path exception

2017-08-01 Thread Chesnay Schepler
One problem i know of is that windows paths with a scheme are not detected as windows paths, as documented in FLINK-6889. They generally still work though (/maybe /by chance). I just verified that calling FileInputFormat#setFilePath() works for both "file:///" and "file:/" on Windows. (I'm assu

SQL API with Table API

2017-08-01 Thread nragon
Hi, Can i expect the output from this: Table revenue = tableEnv.sql( "SELECT TUMBLE_START(EventTime, INTERVAL '30' MINUTE) as tStart, " + "TUMBLE_END(EventTime, INTERVAL '30' MINUTE) as tEnd, " + "cID, " + "cName, " +

multiple users per flink deployment

2017-08-01 Thread Georg Heiler
Hi, flink currently only seems to support a single kerberos ticket for deployment. Are there plans to support different users per each job? regards, Georg

Re: Flink CLI cannot submit job to Flink on Mesos

2017-08-01 Thread Francisco Gonzalez Barea
Hey! It´s working now!! I will do a summary for those who might have the same problem in the future: - Flink 1.3.0 dockerized on Mesos: - Add the HA configuration values in your flink app: https://ci.apache.org/projects/flink/flink-docs-release-1.3/setup/jobmanager_high_availability.html#config-

Re: Access Sliding window

2017-08-01 Thread Fabian Hueske
The average would be computed over the aggregated 15-minute count values. The sliding window would emit every 15 minutes the average of all records that arrived within the last 6 hours. Since the preceding 15-minute tumbling window emits 1 record every 15 mins, this would be the avg over 24 records

Re: Using FileInputFormat - org.apache.flink.streaming.api.functions.source.TimestampedFileInputSplit

2017-08-01 Thread Fabian Hueske
Hi Mohit, these are just INFO log statements that do not necessarily indicate a problem. Is the program working otherwise or do you observe other problems? Best, Fabian 2017-08-01 0:32 GMT+02:00 Mohit Anchlia : > I even tried existing format but still same error: > > FileInputFormat fileInputFo

Re: Does Flink support long term storage and complex queries

2017-08-01 Thread Fabian Hueske
Hi Basanth, Flink is not a storage system (neither for stream nor for batch data). Flink applications can be stateful and maintain very large state (in the order of several TBs) but state is always associated with an application. State can be queryable, so outside applications can run key-lookup (

Re: Odd flink behaviour

2017-08-01 Thread Fabian Hueske
An InputFormat processes multiple InputSplits. open() is called for each InputSplit. If you don't reset reached to false in open() you will only read a single (i.e., the first) InputSplit and skip all others. I'd override open as follows: public void open(FileInputSplit fileSplit) throws IOExcept

Re: KeyBy State

2017-08-01 Thread Fabian Hueske
Hi, there is no built-in support for key changes. You might be able to feedback a changed key with an iteration edge, but not sure how well that works. Best, Fabian 2017-08-01 7:32 GMT+02:00 Govindarajan Srinivasaraghavan < govindragh...@gmail.com>: > Hi, > > I have a keyby state but the key ca

Re: Flatbuffers and Flink

2017-08-01 Thread Fabian Hueske
Hi, in principle you can use any data type with Flink including byte[]. However, all of your functions need the logic to interpret the bytes and you have to implement custom key extractors (if you need to keyBy or partition your stream). Best, Fabian 2017-08-01 2:09 GMT+02:00 Basanth Gowda : >

Re: Flink CLI cannot submit job to Flink on Mesos

2017-08-01 Thread Francisco Gonzalez Barea
Hi Stephan, So, do you mean to remove the “-m” param from the flink CLI call? And on the other hand, that I should add the Zookeeper configuration in both sides, the remote flink and locally in the flink CLI config, right? Regards On 31 Jul 2017, at 22:21, Stephan Ewen mailto:se...@apache.or

Re: multiple streams with multiple actions - proper way?

2017-08-01 Thread Fabian Hueske
Hi Peter, this kind of use case is supported, but it is best practice to split independent pipelines into individual jobs. One reason for that is to isolate failures and restarts. For example, I would split the program you posted into two programs, one for the "foo" topic and one of the "bar" topi