Re: Submit high version compiled code jar to low version flink cluster?

2019-12-29 Thread tison
It possibly fails with incompatibility. Flink doesn't promise such compatibility but it MIGHT work. Best, tison. wangl...@geekplus.com.cn 于2019年12月30日周一 下午3:17写道: > > The flink cluster version is 1.8.2 > The application source code needs some feature only supported in 1.9.1. > So it is

Re: Error:java: 无效的标记: --add-exports=java.base/sun.net.util=ALL-UNNAMED

2019-12-29 Thread Jark Wu
I have the same problem when upgrade my IDEA to 2019.3.1. And the Biao's solution works for me too. Best, Jark On Thu, 21 Nov 2019 at 22:23, 刘建刚 wrote: > Thank you very much. It works for me. > > > 在 2019年11月14日,下午1:06,Biao Liu 写道: > > > > Hi, > > > > I have encountered the same issue when

Submit high version compiled code jar to low version flink cluster?

2019-12-29 Thread wangl...@geekplus.com.cn
The flink cluster version is 1.8.2 The application source code needs some feature only supported in 1.9.1. So it is compiled with flink-1.9.1 denendency and builed to a fat jar with all the flink dependencies. What it will happen if I submit the high version builed jar to the low verion flink

An issue with low-throughput on Flink 1.8.3 running Yahoo streaming benchmarks

2019-12-29 Thread Shinhyung Yang
Dear Flink Users, I'm running the Yahoo streaming benchmarks (the original version) [1] on Flink 1.8.3 and got 60K tuples per second. Because I got 282K tuples per second with Flink 1.1.3, I would like to ask your opinions where I should look at. I have been using one node for a JobManager and

Re: Setting writeTimestampToKafka from Kafka table descriptor

2019-12-29 Thread Jingsong Li
Hi Steve, There are some discussion in [1], this has been considered, but it is not supported in the current version. >From Fabian's word: > I think timestamp fields of source-sink tables should be handled as follows when emitting the table: - proc-time: ignore - from-field: simply write

Re: Duplicate tasks for the same query

2019-12-29 Thread Jingsong Li
Hi RKandoji, FYI: Blink-planner subplan reusing: [1] 1.9 available. Join Join / \ / \ Filter1 Filter2 Filter1 Filter2 ||=> \ / Project1 Project2Project1 ||

Re: StreamTableEnvironment.registerDatastream() 开放用户自定义的schemaDescriptionh和DeserializationSchema

2019-12-29 Thread Terry Wang
你这种需求的一种解决思路,可以把 JSONObject类型定义成object类型,然后注册成table之后通过一个UDTF把JSONObject转换成特定的schema。 Best, Terry Wang > 2019年12月27日 19:56,aven.wu 写道: > > StreamTableEnvironment.registerDatastream(),目前只支持pojo 或者 >

Re: Duplicate tasks for the same query

2019-12-29 Thread Terry Wang
Hi RKandoji~ Could you provide more info about your poc environment? Stream or batch? Flink planner or blink planner? AFAIK, blink planner has done some optimization to deal such duplicate task for one same query. You can have a try with blink planner :

Re: What is the suggested way to validate SQL?

2019-12-29 Thread Jingsong Li
Hi Kaibo, > Validate SQL syntax not need to depend on connector jar At present, sql function strongly need jar dependent support too , but the overall approach is still under discussion, and there is no clear plan at present. But you are right, it really important for platform users. Another way

Re: Exactly-once ambiguities

2019-12-29 Thread Yun Tang
Hi Mohammad I expected to find a description of a mechanism for detecting and ignoring duplicate events in the documentation, although I got the two-phase commit protocol issuing something utterly different. Flink would not detect and ignore duplicate events when processing them but ensure

Setting writeTimestampToKafka from Kafka table descriptor

2019-12-29 Thread Steve Whelan
Examining the *org.apache.flink.table.descriptors.Kafka* class in Flink v1.9, it seems to not have the ability to set whether the Kafka producer should attach a timestamp to the message. The *FlinkKafkaProducer* class has a setter for controlling this producer attribute. Can/should this attribute

Re: Stateful function metrics

2019-12-29 Thread Igal Shilman
Hi Dan, You can learn more about Flink’s metrics system at [1] You would be able to either setup a reporter that would export the metrics to an external system, or query the metrics via the REST API, or simply use Flink’s web ui to obtain them. If I understand the second part of your question

Fwd: Duplicate tasks for the same query

2019-12-29 Thread RKandoji
Hi Team, I'm doing a POC with flink to understand if it's a good fit for my use case. As part of the process, I need to filter duplicate items and created below query to get only the latest records based on timestamp. For instance, I have "Users" table which may contain multiple messages for the

Exactly-once ambiguities

2019-12-29 Thread Mohammad NM
Dear Flink team, I have some ambiguity when it comes to Flink's exactly-once guaranteeing. 1. Based on what I understand, when a failure occurs, some events will be replayed which causes them to appear twice in the computations. I cannot realize how the two-phase commit protocol can avoid this

Stateful function metrics

2019-12-29 Thread Dan Pettersson
Hi all I'm trying to get hold of some metrics from the functions that I have created but can't find a way to get them. It's the metrics mentioned here I'm interested about: https://statefun.io/deployment_operations/metrics.html Any suggestions are appreciated. I also have a question regarding

[ANNOUNCE] Weekly Community Update 2019/52

2019-12-29 Thread Hequn Cheng
Dear community, Happy to share a short community update this week. Due to the holiday, the dev@ mailing list is pretty quiet these days. Flink Development == * [sql] Jark proposes to correct the terminology of "Time-windowed Join" to "Interval Join" in Table API & SQL before 1.10 is

Re: [QUESTION] Why does Apache Flink add `org.antlr` to pom.xml

2019-12-29 Thread Juan Pan
Hi Benchao, Thanks for your reply, key word `antlr` appeared in three modules of master branch, 1. flink-python 2. flink-sql-client 3. flink-hive-connector I also found just tag `exclusion` is used for `Antlr`, so the aim of which is to exclude it from 'hive-exec’ dependency, right? That