Ok, thanks Fabian.
From: Fabian Hueske [mailto:fhue...@gmail.com]
Sent: Tuesday, October 11, 2016 1:12 AM
To: user@flink.apache.org
Subject: Re: Wordindex conversation.
Hi,
you can do it like this:
1) you have to split each label record of the main dataset into separate
records:
Hi,
you can do it like this:
1) you have to split each label record of the main dataset into separate
records:
(0,List(a, b, c, d, e, f, g)) -> (0, a), (0, b), (0, c), ..., (0, g)
(1,List(b, c, f, a, g)) -> (1, b), (1, c), ..., (1, g)
2) join word index dataset with splitted main dataset:
Hi;
I have MainDataset (Label,WordList) :
(0,List(a, b, c, d, e, f, g))
(1,List(b, c, f, a, g))
..and, wordIndex dataset(created with .zipWithIndex) :
wordIndex> (0,a)
wordIndex> (1,b)
wordIndex> (2,c)
wordIndex> (3,d)
wordIndex> (4,e)
wordIndex> (5,f)
wordIndex> (6,g)
Thaks Stephan for the answer.
As I told to Fabian we need to apply some transformation to datasets
interactively.
For the moment I will use livy + spark[1] but I'll prefer to stick with
Flink if possible.
So, if there's any effor in this direction just let me know and I'll be
happy to contribute.
There is still quite a bit needed to do this properly:
(1) incremental recovery
(2) network stack caching
(1) will probably happen quite soon, I am not aware of any committer having
concrete plans for (2).
Best,
Stephan
On Sat, Oct 8, 2016 at 4:41 PM, Flavio Pompermaier
Whoops, I meant Flink and not Beam (I had just visited the Beam mailing list).
-Max
On Mon, Oct 10, 2016 at 12:08 PM, Maximilian Michels wrote:
> Normally, you should be able to directly execute your Beam program
> from within your IDE. It automatically starts a local cluster
Normally, you should be able to directly execute your Beam program
from within your IDE. It automatically starts a local cluster with the
resources needed for the job.
Which Beam version are you using? Could you post some of the code your
executing?
-Max
On Sat, Oct 8, 2016 at 7:51 PM, Dayong
It's from Jun and Unassigned :(
Is There a Workarround?
I'm will try to contact with the reporter , Martin Scholl )
2016-10-10 11:04 GMT+02:00 Timo Walther :
> I think you already found the correct issue describing your problem (
> FLINK-4108). This should get higher
I think you already found the correct issue describing your problem (
FLINK-4108). This should get higher priority.
Timo
Am 09/10/16 um 13:27 schrieb Alberto Ramón:
After solved some issues, I connected with Kylin, but I can't read data
import org.apache.flink.api.scala._
import
Thank you Fabian and Stephan for the suggestions.
I couldn't override "readLine()" because it's final, so went with Fabian's
solution, but I'm struggling with csv field masks. Any help is appreciated.
I created an Input Format which is basically TupleCsvInputFormat for which
I overrode the
10 matches
Mail list logo