Re: Confusion about multiple use of one ValueState

2016-05-13 Thread Balaji Rajagopalan
I wrote a simple helper class, the redis connection are initialized in the constructor and there are set and get methods to store and retreive values from your map functions. If you find any better way to do this please share :). I am using redis scala client. object class RedisHelper { val re

Re: Confusion about multiple use of one ValueState

2016-05-13 Thread Balaji Rajagopalan
Even thought there are multiple instance of map object transient value object state is accessible across the object, so as the stream is flowing in the value can be updated based on application logic. On Fri, May 13, 2016 at 11:26 AM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com>

Re: Confusion about multiple use of one ValueState

2016-05-12 Thread Balaji Rajagopalan
I don't think the valuestate defined in one map function is accessible in other map function this is my understanding, also you need to be aware there will be instance of map function created for each of your tuple in your stream, I had a similar use case where I had to pass in some state from one

Re: reading from latest kafka offset when flink starts

2016-05-12 Thread Balaji Rajagopalan
expected. On Wed, May 11, 2016 at 11:42 PM, Aljoscha Krettek wrote: > Hi, > are you per change using Kafka 0.9? > > Cheers, > Aljoscha > > On Tue, 10 May 2016 at 08:37 Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> Robert, >> Re

Re: reading from latest kafka offset when flink starts

2016-05-09 Thread Balaji Rajagopalan
ailing list > already. > > – Ufuk > > On Fri, May 6, 2016 at 12:07 PM, Balaji Rajagopalan > wrote: > > Thanks Robert appreciate your help. > > > > On Fri, May 6, 2016 at 3:07 PM, Robert Metzger > wrote: > >> > >> Hi, > >>

question regarding windowed stream

2016-05-06 Thread Balaji Rajagopalan
I have a requirement where I want to do aggregation on one data stream every 5 minutes, a different data stream every 1 minute. I wrote a example code to test this out but the behavior is different from what I expected , I expected the window2 to be called 5 times, and window 1 to called once , but

Re: reading from latest kafka offset when flink starts

2016-05-06 Thread Balaji Rajagopalan
to the current offset ("earliest" is the opposite). > > How heavy is the processing you are doing? 4500 events/second sounds not > like a lot of throughput. > > On Wed, May 4, 2016 at 8:24 AM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> I

Re: Flink - start-cluster.sh

2016-05-03 Thread Balaji Rajagopalan
What is the flink documentation you were following to set up your cluster , can you point to that ? On Tue, May 3, 2016 at 6:21 PM, Punit Naik wrote: > Hi > > I did all the settings required for cluster setup. but when I ran the > start-cluster.sh script, it only started one jobmanager on the ma

reading from latest kafka offset when flink starts

2016-05-03 Thread Balaji Rajagopalan
I am using the flink connector to read from a kafka stream, I ran into the problem where the flink job went down due to some application error, it was down for sometime, meanwhile the kafka queue was growing as expected no consumer to consume from the given group , and when I started the flink it s

Re: logback.xml and logback-yarn.xml rollingpolicy configuration

2016-04-21 Thread Balaji Rajagopalan
gt; Cheers, > Till > > On Tue, Apr 19, 2016 at 2:01 PM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> The are two files in the /usr/share/flink/conf directory, and I was >> trying to do the rolling of application logs which goes to following >

Re: Leader not found

2016-04-20 Thread Balaji Rajagopalan
Leader: 0 Replicas: 0 Isr: 0 On Wed, Apr 20, 2016 at 10:35 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com> wrote: > /usr/share/kafka_2.11-0.8.2.1/bin$ ./kafka-topics.sh --describe --topic > capi --zookeeper (someserver) > > Topic:capi PartitionCount:1 Replicati

Re: Leader not found

2016-04-20 Thread Balaji Rajagopalan
.0 and 0.8.2.1 and for both versions > everything worked fine. > How many partitions does your topic have? > > Can you send me the full logs of the Kafka consumer? > > On Tue, Apr 19, 2016 at 6:05 PM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >&

Re: Leader not found

2016-04-19 Thread Balaji Rajagopalan
issue? > > On Tue, Apr 19, 2016 at 2:06 PM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> It does not seem to fully work if there is no data in the kafka stream, >> the flink application emits this error and bails, could this be missed use >

Re: ClasNotFound when submitting job from command line

2016-04-19 Thread Balaji Rajagopalan
In your pom.xml add the maven.plugins like this, and you will have to add all the dependent artifacts, this works for me, if you fire mvn clean compile package, the created jar is a fat jar. org.apache.maven.plugins maven-dependency-plugin 2.9

Re: Leader not found

2016-04-19 Thread Balaji Rajagopalan
fail on a leader > change, then Flink will restart the Kafka consumers and they'll find the > new leaders. > Starting from Flink 1.0.0 the Kafka consumer will handle leader changes > without failing. > > Regards, > Robert > > On Tue, Apr 19, 2016 at 12:17 PM, Balaji R

logback.xml and logback-yarn.xml rollingpolicy configuration

2016-04-19 Thread Balaji Rajagopalan
The are two files in the /usr/share/flink/conf directory, and I was trying to do the rolling of application logs which goes to following directory in task nodes. /var/log/hadoop-yarn/containers/application_*/container_*/taskmanager.log out err Changing the logback.xml and logback-yarn.xml has no

Leader not found

2016-04-19 Thread Balaji Rajagopalan
I am facing this exception repeatedly while trying to consume from kafka topic. It seems it was reported in 1.0.0 and fixed in 1.0.0, how can I be sure that is fixed in the version of flink that I am using, does it require me to install patch updates ? Caused by: java.lang.RuntimeException: Unabl

Re: Flink 0.10.2 and Kafka 0.8.1

2016-04-17 Thread Balaji Rajagopalan
I had fought with 0.8.0.2 kafka and flink 0.10.2 scala version 2.11, was never able to get it working confounded with noclassdeffounderror, moved to flink 1.0.0 with kafka 0.8.0.2 scala version 2.11 things worked for me, if moving to flink 1.0.0 is an option for you do so. balaji On Mon, Apr 18,

Re: Configuring task slots and parallelism for single node Maven executed

2016-04-17 Thread Balaji Rajagopalan
Answered based on my understanding. On Mon, Apr 18, 2016 at 8:12 AM, Prez Cannady wrote: > Some background. > > I’m running Flink application on a single machine, instrumented by Spring > Boot and launched via the Maven Spring Boot plugin. Basically, I’m trying > to figure out how much I can squ

Re: OOME PermGen in URLClassLoader

2016-04-15 Thread Balaji Rajagopalan
Not a solution for your problem,but an alternative, I wrote my own sink function where I handle all sql activities(insert/update/select), used a 3rd lib for connection pooling, the code has been running stable in production without any issue. On Fri, Apr 15, 2016 at 1:41 PM, Maximilian Bode < maxi

Re: streaming join implementation

2016-04-13 Thread Balaji Rajagopalan
the non-matching records after 5 > minutes into redis? Does flink give you control on which records is not > matching in the current window such that you can copy into a long-term > storage? > > > > On Wed, Apr 13, 2016 at 11:20 PM, Balaji Rajagopalan < > balaji.rajagopa..

Re: streaming join implementation

2016-04-13 Thread Balaji Rajagopalan
You can implement join in flink (which is a inner join) the below mentioned pseudo code . The below join is for a 5 minute interval, yes will be some corners cases when the data coming after 5 minutes will be missed out in the join window, I actually had solved this problem but storing some data i

Re: RemoteTransportException when trying to redis in flink code

2016-04-06 Thread Balaji Rajagopalan
cause in the > stack trace it's written that you access a RedisClientPool from the > DriverStreamHelper.set method. > > Cheers, > Till > > > On Wed, Apr 6, 2016 at 11:42 AM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> Till, >&g

Re: RemoteTransportException when trying to redis in flink code

2016-04-06 Thread Balaji Rajagopalan
ure here. } } On Wed, Apr 6, 2016 at 2:42 PM, Till Rohrmann wrote: > Hi Balaji, > > from the stack trace it looks as if you cannot open a connection redis. > Have you checked that you can access redis from all your TaskManager nodes? > > Cheers, > Till > > On Wed, Apr

RemoteTransportException when trying to redis in flink code

2016-04-05 Thread Balaji Rajagopalan
I am trying to use AWS EMR yarn cluster where the flink code runs, in one of apply window function, I try to set some values in redis it fails. I have tried to access the same redis with no flink code and get/set works, but from the flink I get into this exception. Any inputs on what might be goin

Re: building for Scala 2.11

2016-04-04 Thread Balaji Rajagopalan
In your pom file you can mention against which version of scala you want to build and also remember to add the scala version in the artifactId in all the dependencies which takes scala version, there might be some libraries which are scala agnostic there you do not have to specify the scala version

Re: FYI: Added Documentation on Basic Concepts

2016-04-04 Thread Balaji Rajagopalan
Nice write up, one question though my understanding of keyed stream is that it will fork n number of streams from one stream based on n keys, if that is true it can be pictorially depicted as well and the apply function will can be shown to operate over the time period by clearly marking a time lin

Re: Kafka Test Error

2016-04-01 Thread Balaji Rajagopalan
gt; 'org.apache.kafka:kafka-clients:0.9.0.0', >> 'org.apache.flink:flink-java:1.0.0', >> 'org.apache.flink:flink-streaming-java_2.10:1.0.0', >> 'org.apache.flink:flink-connector-kafka-0.9_2.10:1.0.0', >> 'org.apache.flink:flink-connector

Re: Multiple operations on a WindowedStream

2016-04-01 Thread Balaji Rajagopalan
I had a similar use case and ended writing the aggregation logic in the apply function, could not find any better solution. On Fri, Apr 1, 2016 at 6:03 AM, Kanak Biscuitwala wrote: > Hi, > > I would like to write something that does something like a word count, and > then emits only the 10 highe

Re: Kafka Test Error

2016-03-30 Thread Balaji Rajagopalan
I have tested kafka 0.8.0.2 with flink 1.0.0 and it works for me. Can't talk about kafka 0.9.0.1. On Wed, Mar 30, 2016 at 12:51 PM, Zhun Shen wrote: > Hi there, > > flink version: 1.0.0 > kafka version: 0.9.0.0 > env: local > > I run the script below: > ./bin/flink run -c com.test.flink.FlinkTes

Re: jackson DefaultScalaModule missing on deployment

2016-03-29 Thread Balaji Rajagopalan
You will have to include dependent jackson jar in flink server lib folder, or create a fat jar. balaji On Tue, Mar 29, 2016 at 4:47 PM, Bart van Deenen wrote: > Hi all > > I've succesfully built a Flink streaming job, and it runs beautifully in > my IntelliJ ide, with a Flink instance started o

Re: flink streaming - window chaining example

2016-03-27 Thread Balaji Rajagopalan
val stream:DataStream[String] = env .addSource(new FlinkKafkaConsumer08[String]("topic_name", new SimpleStringSchema, prop)) val event:DataStream[SomeEventObj] = stream.map(MyMapFunction) val tenMinute:DataStream[AggEvents] = ridesByDeviceIdStream.timeWindowAll(Time.of(10, TimeUnit.MINUTES).tri

Re: does reduce function has a bug

2016-03-25 Thread Balaji Rajagopalan
Never mind Till figured out a way, instead of doing the aggregation in reduce, I moved that logic to apply of the window function. On Thu, Mar 24, 2016 at 11:33 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com> wrote: > Till, > > Thanks for your reply, may be I should

Re: does reduce function has a bug

2016-03-24 Thread Balaji Rajagopalan
ractical, Flink emits the partial > reduce results. > > Cheers, > Till > > On Thu, Mar 24, 2016 at 6:21 PM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> I have keyed input stream on DateStream(String,Int) and wrote a reduce on >> the ke

does reduce function has a bug

2016-03-24 Thread Balaji Rajagopalan
I have keyed input stream on DateStream(String,Int) and wrote a reduce on the keyedStream. The reduce is simple one summing up the integer values of the same key. val stream = DataStream(String,Int) val keyedStream = stream.keyBy(_._1).reduce( new MyReduceFunction) keyedStream.print() class MyRed

Re: Valuestate is not saving the state

2016-03-23 Thread Balaji Rajagopalan
Never mind I understand what is going on Aljoscha for each unique key the value count is reset to 0. On Wed, Mar 23, 2016 at 4:37 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com> wrote: > (Booking(te7uc4,compact,j...@gmail.com,Mon Feb 29 19:19:40 IST > 2016),145873098,1

Re: Valuestate is not saving the state

2016-03-23 Thread Balaji Rajagopalan
> > Cheers, > Aljoscha > > On 23 Mar 2016, at 11:53, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > > > > I wrote the below code which will increment a counter for the data in > the datastream, and when I print the counter each time it seems the

Valuestate is not saving the state

2016-03-23 Thread Balaji Rajagopalan
I wrote the below code which will increment a counter for the data in the datastream, and when I print the counter each time it seems the value is reinitialised to 0, and it is not incrementing, any thoughts. class BookingCntFlatMapFunction extends RichFlatMapFunction[(Booking,Long,Long),(Booking,

Re: Access to S3 from YARN on EC2

2016-03-21 Thread Balaji Rajagopalan
This kind of class not found exception is a little bit misleading, it is not the class is not found is the real problem rather than the combination of the different libraries that are using there is a version compatibility mismatch, so you will have to go back and check if there is any version mism

Re: Help with DeltaIteration

2016-03-19 Thread Balaji Rajagopalan
The easier way to debug this would be have prints in the projectjoinresultmapper and see what data you are getting. It is possible your original dataset has duplicate rows ? On Thu, Mar 17, 2016 at 6:36 PM, Lydia Ickler wrote: > Hi, > I have a question regarding the Delta Iteration. > I basicall

Re: How to start with the first Kafka Message

2016-03-18 Thread Balaji Rajagopalan
If it is a one time thing, you could run the kafka consumer script which has the --from-beginning option, and redirect the output as a socketstream and consume in flink. On Thu, Mar 17, 2016 at 7:24 PM, Till Rohrmann wrote: > Hi Dominique, > > have you tried setting the Kafka property props.put(

Re: how to convert DataStream to DataSet

2016-03-15 Thread Balaji Rajagopalan
There was a similar question before the answer was to use org.apache.flink.api.common.io.OutputFormat to do the conversion. On Tue, Mar 15, 2016 at 7:48 PM, subash basnet wrote: > Hello all, > > In WikipediaAnalysis.java we get *result *of type *DataStream Long>>*, > > I would want to convert *r

Re: kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-15 Thread Balaji Rajagopalan
Robert, I got it working for 1.0.0. balaji On Mon, Mar 14, 2016 at 4:41 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com> wrote: > Yep the same issue as before(class not found) with flink 0.10.2 with > scala version 2.11. I was not able to use scala 2.10 since c

Re: kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-14 Thread Balaji Rajagopalan
Yep the same issue as before(class not found) with flink 0.10.2 with scala version 2.11. I was not able to use scala 2.10 since connector for flink_connector_kafka for 0.10.2 is not available. balaji On Mon, Mar 14, 2016 at 4:20 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com>

Re: kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-14 Thread Balaji Rajagopalan
or-kafka_ doesn't exist for 1.0.0. You have to use either > flink-connector-kafka-0.8_ or flink-connector-kafka-0.9_ > > > On Mon, Mar 14, 2016 at 11:17 AM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> What I noticied was that, if I remove the dep

Re: kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-14 Thread Balaji Rajagopalan
What I noticied was that, if I remove the dependency on flink-connector-kafka so it is clearly to do something with that dependency. On Mon, Mar 14, 2016 at 3:46 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com> wrote: > Robert, >I have moved on to latest version of fl

Re: kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-14 Thread Balaji Rajagopalan
n you send me the full build file to further investigate the issue? > > On Fri, Mar 11, 2016 at 4:56 PM, Balaji Rajagopalan < > balaji.rajagopa...@olacabs.com> wrote: > >> Robert, >> That did not fix it ( using flink and connector same version) . Tried >>

Re: kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-11 Thread Balaji Rajagopalan
from the > "org.apache.flink" group. > > You said these are the versions you are using: > > flink.version = 0.10.2 > kafka.verison = 0.8.2 > flink.kafka.connection.verion=0.9.1 > > For the connector, you also need to use 0.10.2. > > > > On Fri, Mar 1

kafka.javaapi.consumer.SimpleConsumer class not found

2016-03-11 Thread Balaji Rajagopalan
I am tyring to use the flink kafka connector, for this I have specified the kafka connector dependency and created a fat jar since default flink installation does not contain kafka connector jars. I have made sure that flink-streaming-demo-0.1.jar has the kafka.javaapi.consumer.SimpleConsumer.class

Re: 404 error for Flink examples

2016-03-10 Thread Balaji Rajagopalan
You could try this link. https://ci.apache.org/projects/flink/flink-docs-master/apis/batch/examples.html On Fri, Mar 11, 2016 at 9:56 AM, janardhan shetty wrote: > Hi, > > I was looking at the examples for Flink applications and the comment in > quickstart/job results in 404 for the web page. >

Re: DataSet -> DataStream

2016-03-10 Thread Balaji Rajagopalan
You could I suppose write the dateset to a sink a file and then read the file to a data stream. On Fri, Mar 11, 2016 at 4:18 AM, Prez Cannady wrote: > > I’d like to pour some data I’ve collected into a DataSet via JDBC into a > Kafka topic, but I think I need to transform my DataSet into a DataS

readTextFile is not working for StreamExecutionEnvironment

2016-03-02 Thread Balaji Rajagopalan
def main(args: Array[String]): Unit = { val env: StreamExecutionEnvironment = StreamExecutionEnvironment.getExecutionEnvironment(); env.setStreamTimeCharacteristic(TimeCharacteristic.EventTime) try { val data1:DataStream[String] = env.readTextFile(“somefile.txt”); data1.print()