Hi,
I am running a program that works fine locally, but when I try to run it on the
cluster I get a timeout error from the client that tries to connect to the
jobmanager. There is no issue with contacting the jobmanager form the machine,
as it works just fine for other stream applications. I su
I would try to do the outlier compuation with the DataSet API instead of
fetching the results to the client with collect().
If you do that, you can directly use writeAsCsv because the result is still
a DataSet.
What you have to do, is to translate your findOutliers method into DataSet
API code.
B
Hello Fabian,
As written before code:
*DataSet fElements =
env.fromCollection(findOutliers(clusteredPoints,
finalCentroids));fElements.writeAsCsv(outputPath, "\n", "
");env.execute("KMeans Example");*
I am very new to flink so not so clear about what you suggested, by
option(1) you meant that I
We discovered yesterday that the snapshot builds were not updated in a
while (because the build server experienced timeouts).
Hence the SNAPSHOT build may have quite stale.
It is updating frequently again now, that's probably why you find a correct
build today...
On Wed, Feb 10, 2016 at 5:31 PM
Hi Chiwan, Max,
Thanks for checking. I also downloaded it now and verified the 2.10 jar is
gone :)
A new build must have overwrote yesterday's and corrected itself.
flink-1.0-SNAPSHOT-bin-hadoop2_2.11.tgz
2016-02-10T15:55:33.000Z
Thanks!
David
On Wed, Feb 10, 2016 at 4:44 AM, Maximilian Michel
Hello,
the compiler has been raising an error since I added this line to the code
val testData=streamEnv.addSource(new
FlinkKafkaConsumer082[String]("data-input",new
SimpleStringSchema(),kafkaProp))
Here is the error:
Error:scalac: Class
org.apache.flink.streaming.api.checkpoint.CheckpointN
Hello Fabian,
Thank you very much for the resource. I had already gone through this and
have found port '6123' as default for taskmanager registration. But I want
to know the specific range of ports the taskmanager access during job
execution.
The taskmanager always tries to access a random port
Yes, 0.10.x does not always have Scala version suffixes.
1.0 is doing this consistently, should cause less confusion...
On Wed, Feb 10, 2016 at 2:38 PM, shotte wrote:
> Ok It is working now
>
> I had to change a few dependency with the _2.11 suffix
>
> Thanks
>
> Sylvain
>
>
>
>
Note that some of these config options are only available starting from
version 1.0-SNAPSHOT
On Wed, Feb 10, 2016 at 2:16 PM, Fabian Hueske wrote:
> Hi Ravinder,
>
> please have a look at the configuration documentation:
>
> -->
> https://ci.apache.org/projects/flink/flink-docs-master/setup/conf
Ok It is working now
I had to change a few dependency with the _2.11 suffix
Thanks
Sylvain
org.apache.flink
flink-java
${flink.version}
o
Hi Ravinder,
please have a look at the configuration documentation:
-->
https://ci.apache.org/projects/flink/flink-docs-master/setup/config.html#jobmanager-amp-taskmanager
Best, Fabian
2016-02-10 13:55 GMT+01:00 Ravinder Kaur :
> Hello All,
>
> I need to know the range of ports that are being
Hi,
Thanks for your reply, but I am still a bit confuse.
I have downloaded flink-0.10.1-bin-hadoop27-scala_2.11.tgz
and kafka_2.11-0.9.0.0.tgz
I did not install myself Scala
Now tell me if I understand correctly.
Depending on the version of Flink I have (in my case the scala 2.11) I must
spec
Hi Subash,
I would not fetch the data to the client, do the computation there, and
send it back, just for the purpose of writing it to a file.
Either 1) pull the results to the client and write the file from there or
2) compute the outliers in the cluster.
I did not study your code completely, bu
Hello All,
I need to know the range of ports that are being used during the
master/slave communication in the Flink cluster. Also is there a way I can
specify a range of ports, at the slaves, to restrict them to connect to
master only in this range?
Kind Regards,
Ravinder Kaur
On Wed, Feb 3, 20
What do you mean exactly..? Probably I'm missing something here..remember
that I can specify the right subClass only after the last flatMap, after
the first map neither me nor Flink can know the exact subclass of BaseClass
On Wed, Feb 10, 2016 at 12:42 PM, Stephan Ewen wrote:
> Class hierarchies
Class hierarchies should definitely work, even if the base class has no
fields.
They work more efficiently if you register the subclasses at the execution
environment (Flink cannot infer them from the function signatures because
the function signatures only contain the abstract base class).
On We
Hi Subash,
how is findOutliers implemented?
It might be that you mix-up local and cluster computation. All DataSets are
processed in the cluster. Please note the following:
- ExecutionEnvironment.fromCollection() transforms a client local
connection into a DataSet by serializing it and sending it
Because The classes are not related to each other. Do you think it's a good
idea to have something like this?
abstract class BaseClass(){
String someField;
}
class ExtendedClass1 extends BaseClass (){
String someOtherField11;
String someOtherField12;
String someOtherField13;
...
}
c
Hello Stefano,
Yeah the type casting worked, thank you. But not able to print the Dataset
to the file.
The default below code which writes the KMeans points along with their
centroid numbers to the file works fine:
// feed new centroids back into next iteration
DataSet finalCentro
Why not use an abstract base class and N subclasses?
On Wed, Feb 10, 2016 at 10:05 AM, Fabian Hueske wrote:
> Unfortunately, there is no Either<1,...,n>.
> You could implement something like a Tuple3, Option,
> Option>. However, Flink does not provide an Option type (comes with
> Java8). You wou
Hi David,
Just had a check as well. Can't find a 2.10 Jar in the lib folder.
Cheers,
Max
On Wed, Feb 10, 2016 at 6:17 AM, Chiwan Park wrote:
> Hi David,
>
> I just downloaded the "flink-1.0-SNAPSHOT-bin-hadoop2_2.11.tgz” but there is
> no jar compiled with Scala 2.10. Could you check again?
>
Unfortunately, there is no Either<1,...,n>.
You could implement something like a Tuple3, Option,
Option>. However, Flink does not provide an Option type (comes with
Java8). You would need to implement it yourself incl. TypeInfo and
Serializer. You can get some inspiration from the Either type info
Yes, the intermediate dataset I create then join again between themselves.
What I'd need is a Either<1,...,n>. Is that possible to add?
Otherwise I was thinking to generate a Tuple2 and in the
subsequent filter+map/flatMap deserialize only those elements I want to
group togheter (e.g. t.f0=="someEv
Hi Flavio,
I did not completely understand which objects should go where, but here are
some general guidelines:
- early filtering is mostly a good idea (unless evaluating the filter
expression is very expensive)
- you can use a flatMap function to combine a map and a filter
- applying multiple fu
Hi Fabian,
your hint was good! Maven fools me with the dependency management. Now
everything works as expected!
Many many thanks to all of you!
Greets
Dominique
Am 10.02.2016 um 08:45 schrieb Fabian Hueske:
Hi Dominique,
can you check if the versions of the remotely running job manager &
25 matches
Mail list logo