RE: Wordindex conversation.

2016-10-10 Thread Kürşat Kurt
Ok, thanks Fabian. From: Fabian Hueske [mailto:fhue...@gmail.com] Sent: Tuesday, October 11, 2016 1:12 AM To: user@flink.apache.org Subject: Re: Wordindex conversation. Hi, you can do it like this: 1) you have to split each label record of the main dataset into separate records:

Re: Wordindex conversation.

2016-10-10 Thread Fabian Hueske
Hi, you can do it like this: 1) you have to split each label record of the main dataset into separate records: (0,List(a, b, c, d, e, f, g)) -> (0, a), (0, b), (0, c), ..., (0, g) (1,List(b, c, f, a, g)) -> (1, b), (1, c), ..., (1, g) 2) join word index dataset with splitted main dataset:

Wordindex conversation.

2016-10-10 Thread Kürşat Kurt
Hi; I have MainDataset (Label,WordList) : (0,List(a, b, c, d, e, f, g)) (1,List(b, c, f, a, g)) ..and, wordIndex dataset(created with .zipWithIndex) : wordIndex> (0,a) wordIndex> (1,b) wordIndex> (2,c) wordIndex> (3,d) wordIndex> (4,e) wordIndex> (5,f) wordIndex> (6,g)

Re: Iterative queries on Flink

2016-10-10 Thread Flavio Pompermaier
Thaks Stephan for the answer. As I told to Fabian we need to apply some transformation to datasets interactively. For the moment I will use livy + spark[1] but I'll prefer to stick with Flink if possible. So, if there's any effor in this direction just let me know and I'll be happy to contribute.

Re: Iterative queries on Flink

2016-10-10 Thread Stephan Ewen
There is still quite a bit needed to do this properly: (1) incremental recovery (2) network stack caching (1) will probably happen quite soon, I am not aware of any committer having concrete plans for (2). Best, Stephan On Sat, Oct 8, 2016 at 4:41 PM, Flavio Pompermaier

Re: Executing a flink program

2016-10-10 Thread Maximilian Michels
Whoops, I meant Flink and not Beam (I had just visited the Beam mailing list). -Max On Mon, Oct 10, 2016 at 12:08 PM, Maximilian Michels wrote: > Normally, you should be able to directly execute your Beam program > from within your IDE. It automatically starts a local cluster

Re: Executing a flink program

2016-10-10 Thread Maximilian Michels
Normally, you should be able to directly execute your Beam program from within your IDE. It automatically starts a local cluster with the resources needed for the job. Which Beam version are you using? Could you post some of the code your executing? -Max On Sat, Oct 8, 2016 at 7:51 PM, Dayong

Re: jdbc.JDBCInputFormat

2016-10-10 Thread Alberto Ramón
It's from Jun and Unassigned :( Is There a Workarround? I'm will try to contact with the reporter , Martin Scholl ) 2016-10-10 11:04 GMT+02:00 Timo Walther : > I think you already found the correct issue describing your problem ( > FLINK-4108). This should get higher

Re: jdbc.JDBCInputFormat

2016-10-10 Thread Timo Walther
I think you already found the correct issue describing your problem ( FLINK-4108). This should get higher priority. Timo Am 09/10/16 um 13:27 schrieb Alberto Ramón: After solved some issues, I connected with Kylin, but I can't read data import org.apache.flink.api.scala._ import

Re: Handling decompression exceptions

2016-10-10 Thread Yassine MARZOUGUI
Thank you Fabian and Stephan for the suggestions. I couldn't override "readLine()" because it's final, so went with Fabian's solution, but I'm struggling with csv field masks. Any help is appreciated. I created an Input Format which is basically TupleCsvInputFormat for which I overrode the