Hi Fanbin,

you could use the RC1 of Flink that was created yesterday and use the
apache repo
https://repository.apache.org/content/repositories/orgapacheflink-1325/org/apache/flink/flink-json/1.10.0/
.
Alternatively, if you build Flink locally with `mvn install`, then you
could use mavenLocal() in your gradle.build and feed from that.

Best,

Arvid

On Tue, Jan 28, 2020 at 1:24 AM Fanbin Bu <fanbin...@coinbase.com> wrote:

> I can build flink 1.10 and install it on to EMR
> (flink-dist_2.11-1.10.0.jar). but what about other dependencies in my
> project build.gradle, ie. flink-scala_2.11, flink-json, flink-jdbc... do I
> continue to use 1.9.0 since there is no 1.10 available?
>
> Thanks,
> Fanbin
>
> On Fri, Jan 24, 2020 at 11:39 PM Bowen Li <bowenl...@gmail.com> wrote:
>
>> Hi Fanbin,
>>
>> You can install your own Flink build in AWS EMR, and it frees you from
>> Emr’s release cycles
>>
>> On Thu, Jan 23, 2020 at 03:36 Jingsong Li <jingsongl...@gmail.com> wrote:
>>
>>> Fanbin,
>>>
>>> I have no idea now, can you created a JIRA to track it? You can describe
>>> complete SQL and some data informations.
>>>
>>> Best,
>>> Jingsong Lee
>>>
>>> On Thu, Jan 23, 2020 at 4:13 PM Fanbin Bu <fanbin...@coinbase.com>
>>> wrote:
>>>
>>>> Jingsong,
>>>>
>>>> Do you have any suggestions to debug the above mentioned
>>>> IndexOutOfBoundsException error?
>>>> Thanks,
>>>>
>>>> Fanbin
>>>>
>>>> On Wed, Jan 22, 2020 at 11:07 PM Fanbin Bu <fanbin...@coinbase.com>
>>>> wrote:
>>>>
>>>>> I got the following error when running another job. any suggestions?
>>>>>
>>>>> Caused by: java.lang.IndexOutOfBoundsException
>>>>> at
>>>>> org.apache.flink.core.memory.MemorySegment.getInt(MemorySegment.java:701)
>>>>> at
>>>>> org.apache.flink.table.dataformat.BinaryRow.getInt(BinaryRow.java:264)
>>>>> at HashWinAggWithKeys$538.endInput(Unknown Source)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.tasks.OperatorChain.endInput(OperatorChain.java:276)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.checkFinished(StreamOneInputProcessor.java:151)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:138)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.performDefaultAction(StreamTask.java:276)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.run(StreamTask.java:298)
>>>>> at
>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:403)
>>>>> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
>>>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
>>>>> at java.lang.Thread.run(Thread.java:748)
>>>>>
>>>>> On Wed, Jan 22, 2020 at 8:57 PM Fanbin Bu <fanbin...@coinbase.com>
>>>>> wrote:
>>>>>
>>>>>> Jingsong,
>>>>>>
>>>>>> I set the config value to be too large. After I changed it to a
>>>>>> smaller number it works now!
>>>>>> thanks you for the help. really appreciate it!
>>>>>>
>>>>>> Fanbin
>>>>>>
>>>>>> On Wed, Jan 22, 2020 at 8:50 PM Jingsong Li <jingsongl...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Fanbin,
>>>>>>>
>>>>>>> Looks like your config is wrong, can you show your config code?
>>>>>>>
>>>>>>> Best,
>>>>>>> Jingsong Lee
>>>>>>>
>>>>>>> On Thu, Jan 23, 2020 at 12:41 PM Fanbin Bu <fanbin...@coinbase.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Jingsong,
>>>>>>>>
>>>>>>>> Great, now i got a different error:
>>>>>>>>
>>>>>>>> java.lang.NullPointerException: Initial Segment may not be null
>>>>>>>>        at 
>>>>>>>> org.apache.flink.runtime.memory.AbstractPagedOutputView.<init>(AbstractPagedOutputView.java:65)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.runtime.io.disk.SimpleCollectingOutputView.<init>(SimpleCollectingOutputView.java:49)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap$RecordArea.<init>(BytesHashMap.java:522)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap.<init>(BytesHashMap.java:190)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap.<init>(BytesHashMap.java:149)
>>>>>>>>        at LocalHashWinAggWithKeys$292.open(Unknown Source)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.openAllOperators(StreamTask.java:529)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:393)
>>>>>>>>        at 
>>>>>>>> org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
>>>>>>>>        at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
>>>>>>>>        at java.lang.Thread.run(Thread.java:748)
>>>>>>>>
>>>>>>>>
>>>>>>>> is there any other config i should add?
>>>>>>>>
>>>>>>>> thanks,
>>>>>>>>
>>>>>>>> Fanbin
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, Jan 22, 2020 at 8:07 PM Fanbin Bu <fanbin...@coinbase.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> you beat me to it.
>>>>>>>>> let's me try that.
>>>>>>>>>
>>>>>>>>> On Wed, Jan 22, 2020 at 7:57 PM Jingsong Li <
>>>>>>>>> jingsongl...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Fanbin,
>>>>>>>>>>
>>>>>>>>>> Document is here:
>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/table/config.html
>>>>>>>>>> NOTE: you need configure this into TableConfig.
>>>>>>>>>>
>>>>>>>>>> Best,
>>>>>>>>>> Jingsong Lee
>>>>>>>>>>
>>>>>>>>>> On Thu, Jan 23, 2020 at 11:54 AM Fanbin Bu <
>>>>>>>>>> fanbin...@coinbase.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Jingsong,
>>>>>>>>>>>
>>>>>>>>>>> Thank you for the response.
>>>>>>>>>>> Since I'm using flink on EMR and the latest version is 1.9 now.
>>>>>>>>>>> the second option is ruled out. but will keep that in mind for 
>>>>>>>>>>> future
>>>>>>>>>>> upgrade.
>>>>>>>>>>>
>>>>>>>>>>> I'm going to try the first option. It's probably a good idea to
>>>>>>>>>>> add that in the doc for example:
>>>>>>>>>>> https://ci.apache.org/projects/flink/flink-docs-stable/ops/config.html
>>>>>>>>>>>
>>>>>>>>>>> Thanks,
>>>>>>>>>>> Fanbin
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Jan 22, 2020 at 7:08 PM Jingsong Li <
>>>>>>>>>>> jingsongl...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Fanbin,
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks for using blink batch mode.
>>>>>>>>>>>>
>>>>>>>>>>>> The OOM is caused by the manage memory not enough in Hash
>>>>>>>>>>>> aggregation.
>>>>>>>>>>>>
>>>>>>>>>>>> There are three options you can choose from:
>>>>>>>>>>>>
>>>>>>>>>>>> 1.Is your version Flink 1.9? 1.9 still use fix memory
>>>>>>>>>>>> configuration. So you need increase hash memory:
>>>>>>>>>>>> - table.exec.resource.hash-agg.memory: 1024 mb
>>>>>>>>>>>>
>>>>>>>>>>>> 2.In 1.10, we use slot manage memory to dynamic config real
>>>>>>>>>>>> operator memory, so operator can use more manage memory, so you 
>>>>>>>>>>>> don't need
>>>>>>>>>>>> configure hash agg memory anymore. You can try 1.10 RC0 [1]
>>>>>>>>>>>>
>>>>>>>>>>>> 3.We can use sort aggregation to avoid OOM too, but there is no
>>>>>>>>>>>> config option now, I created JIRA to track it. [2]
>>>>>>>>>>>>
>>>>>>>>>>>> [1]
>>>>>>>>>>>> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/ANNOUNCE-Apache-Flink-1-10-0-release-candidate-0-td36770.html
>>>>>>>>>>>> [2] https://issues.apache.org/jira/browse/FLINK-15732
>>>>>>>>>>>>
>>>>>>>>>>>> Best,
>>>>>>>>>>>> Jingsong Lee
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Jan 23, 2020 at 7:40 AM Fanbin Bu <
>>>>>>>>>>>> fanbin...@coinbase.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> tried to increase memory:
>>>>>>>>>>>>> flink run  -m yarn-cluster -p 16 -ys 1 -ytm 200000 -yjm 8096 myjar
>>>>>>>>>>>>>
>>>>>>>>>>>>> and still got the same OOM exception.
>>>>>>>>>>>>>
>>>>>>>>>>>>> my sql is like:
>>>>>>>>>>>>>
>>>>>>>>>>>>> select id, hop_end(created_at, interval '30' second, interval '1' 
>>>>>>>>>>>>> minute), sum(field)... #20 of these sums
>>>>>>>>>>>>>
>>>>>>>>>>>>> from table group by id, hop(created_at, interval '30' second, 
>>>>>>>>>>>>> interval '1' minute)
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Wed, Jan 22, 2020 at 3:19 PM Fanbin Bu <
>>>>>>>>>>>>> fanbin...@coinbase.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I have a batch job using blink planner. and got the following
>>>>>>>>>>>>>> error. I was able to successfully run the same job with flink 
>>>>>>>>>>>>>> 1.8 on yarn.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I set conf as:
>>>>>>>>>>>>>> taskmanager.heap.size: 50000m
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> and flink UI gives me
>>>>>>>>>>>>>> Last Heartbeat:20-01-22
>>>>>>>>>>>>>> 14:56:25ID:container_1579720108062_0018_01_000020Data 
>>>>>>>>>>>>>> Port:41029Free Slots
>>>>>>>>>>>>>> / All Slots:1 / 0CPU Cores:16Physical Memory:62.9 GBJVM Heap 
>>>>>>>>>>>>>> Size:10.3
>>>>>>>>>>>>>> GBFlink Managed Memory:24.9 GB
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> any suggestions on how to move forward?
>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>> Fanbin
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Caused by:
>>>>>>>>>>>>>> org.apache.flink.runtime.client.JobExecutionException: Job 
>>>>>>>>>>>>>> execution failed.
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
>>>>>>>>>>>>>> ... 25 more
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *Caused by: java.io.IOException: Hash window aggregate map
>>>>>>>>>>>>>> OOM.* at HashWinAggWithKeys$534.processElement(Unknown
>>>>>>>>>>>>>> Source)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processElement(StreamOneInputProcessor.java:164)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:143)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.performDefaultAction(StreamTask.java:276)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.run(StreamTask.java:298)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:403)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
>>>>>>>>>>>>>> at
>>>>>>>>>>>>>> org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:748)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Best, Jingsong Lee
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best, Jingsong Lee
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best, Jingsong Lee
>>>>>>>
>>>>>>
>>>
>>> --
>>> Best, Jingsong Lee
>>>
>>

Reply via email to