[GitHub] [flink] wuchong commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-24 Thread GitBox
wuchong commented on issue #9181: [FLINK-13210][hive] Hive connector test 
should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-514907840
 
 
   Travis passed: https://travis-ci.com/flink-ci/flink/builds/120541820
   Will merge this.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9211: [FLINK-13393][FLINK-13391][table-planner-blink] Fix source conversion and source return type

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9211: 
[FLINK-13393][FLINK-13391][table-planner-blink] Fix source conversion and 
source return type
URL: https://github.com/apache/flink/pull/9211#issuecomment-514475275
 
 
   ## CI report:
   
   * b465c69b21a3ac810bc34e4f95dc0b3a3d93281c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120311139)
   * 645face305b998b122ee5fb2d731926179885432 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120431665)
   * 7dfc56376b2c4ec37fb38c91851e20e76128ad7c : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120547055)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9222: [FLINK-13409][python] Supported java 
UDFs in python API.
URL: https://github.com/apache/flink/pull/9222#issuecomment-514866924
 
 
   ## CI report:
   
   * 69b88611a023533b59dee72fe56332d77ee93884 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541108)
   * d69fada6a03f2cebbfa4944bed8906cc84e39e79 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120552328)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] dianfu commented on issue #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
dianfu commented on issue #9222: [FLINK-13409][python] Supported java UDFs in 
python API.
URL: https://github.com/apache/flink/pull/9222#issuecomment-514902584
 
 
   The latest PR LGTM. +1 from my side.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9195: [FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert table sink

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9195: [FLINK-13289][table-planner-blink] 
Blink-planner should setKeyFields to upsert table sink
URL: https://github.com/apache/flink/pull/9195#issuecomment-513690205
 
 
   ## CI report:
   
   * 5a275f751d5b7e13cbf4ce2ecacaf923bc8ccaa8 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119973212)
   * e959fd39d8a2d34b92591d2cd31ec405625c85b4 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120119338)
   * 094123574c35c48ce7c93c1e47cd35a81818ad40 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120429194)
   * 4c5fe1ac569d3f4a8ae11f8eeaf4d7090baf425c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120546407)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9223: [FLINK-13352][hive] Using hive connector with hive-1.2.1 needs libfb3…

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9223: [FLINK-13352][hive] Using hive 
connector with hive-1.2.1 needs libfb3…
URL: https://github.com/apache/flink/pull/9223#issuecomment-514873847
 
 
   ## CI report:
   
   * 3ac1a7e5322324b5ce2fd0aaf14cfbecc4793670 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120543108)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-13401) Flink SQL sqlUpdate sql where complex conditional expressions do not take effect

2019-07-24 Thread pingle wang (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13401?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

pingle wang closed FLINK-13401.
---
Resolution: Not A Bug

> Flink SQL sqlUpdate sql where complex conditional expressions do not take 
> effect
> 
>
> Key: FLINK-13401
> URL: https://issues.apache.org/jira/browse/FLINK-13401
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Affects Versions: 1.6.3, 1.7.2
>Reporter: pingle wang
>Priority: Major
> Fix For: 1.9.0
>
> Attachments: image-2019-07-24-16-40-36-070.png, 
> image-2019-07-24-16-45-20-682.png, image-2019-07-24-17-00-36-079.png
>
>
> flink execute sql :
>  
> {code:java}
> select
> count(ad_action) as cnt
> from
> ad_sdk_android
> where
> ad_position_id = '100055'
> and ad_action = 'click'
> and (event_type IS NULL OR event_type <> '2'AND event_type <> '10')
> {code}
> the result is 20, but in mysql and hive, this result is 17.
> !image-2019-07-24-17-00-36-079.png!
>  
> The key condition is :
>  
> {code:java}
> and (event_type IS NULL OR event_type <> '2'AND event_type <> '10'){code}
> and event_type is varchar, flink is string data type.
>  
> mysql has the key condition sql result is:
> !image-2019-07-24-16-40-36-070.png!
> mysql no the key condition sql result is:
> !image-2019-07-24-16-45-20-682.png!
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13413) Put flink-table jar under lib folder

2019-07-24 Thread Rui Li (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892403#comment-16892403
 ] 

Rui Li commented on FLINK-13413:


+1 to put the jar in lib. It surprised me when I run some simple SQL via 
{{sql-client.sh}} and got CNFE that sql parser is not in classpath.

> Put flink-table jar under lib folder
> 
>
> Key: FLINK-13413
> URL: https://issues.apache.org/jira/browse/FLINK-13413
> Project: Flink
>  Issue Type: Improvement
>  Components: Release System
>Affects Versions: 1.9.0
>Reporter: Jeff Zhang
>Priority: Major
>
> Now flink-table jar is in opt folder. Since we plan to make flink-table api 
> as the first citizen of flink, we should put it under lib folder, otherwise 
> it is not convenient to use some tools that use flink table api, such as 
> flink scala shell, sql client and etc. 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9181: [FLINK-13210][hive] Hive connector 
test should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-513213135
 
 
   ## CI report:
   
   * d9311024212f04cca28d1105ab51512c4de4aedd : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119788315)
   * 8a12e7fc07dff371615365cfcbea038a02c232df : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119875298)
   * c4af2954d620a1fc0e8fb84443a6e09a7f258e66 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120364607)
   * a2d681a2a698cc1ff6053afcfd4cdafe392626dd : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541820)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13410) Csv input format does not support LocalDate

2019-07-24 Thread Caizhi Weng (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng updated FLINK-13410:

Issue Type: Sub-task  (was: Bug)
Parent: FLINK-13285

> Csv input format does not support LocalDate
> ---
>
> Key: FLINK-13410
> URL: https://issues.apache.org/jira/browse/FLINK-13410
> Project: Flink
>  Issue Type: Sub-task
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>
> Csv input format is lacking parsers for LocalDate, LocalTime, etc. As 
> DataTypes.DATE now defaults to LocalDate, we should add these parsers for 
> user experience.
> A temporal workaround for the users is that, users can call 
> DataTypes.Timestamp().bridgeTo(java.sql.Timestamp) to use the old 
> SqlTimestamp converter.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot commented on issue #9224: [FLINK-13266] [connectors / kafka] Fix race condition between transaction commit and produc…

2019-07-24 Thread GitBox
flinkbot commented on issue #9224: [FLINK-13266] [connectors / kafka] Fix race 
condition between transaction commit and produc…
URL: https://github.com/apache/flink/pull/9224#issuecomment-514889442
 
 
   ## CI report:
   
   * 78b445e8a5fb88c8c7f7080f1e6d9fdf89a5a594 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120548015)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307108029
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/ExecutionConfigOptions.java
 ##
 @@ -0,0 +1,202 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table module.
+ */
+public class ExecutionConfigOptions {
+
+   // 

+   //  Source Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SOURCE_IDLE_TIMEOUT 
=
+   key("table.exec.source.idle-timeout")
+   .defaultValue("-1 ms")
+   .withDescription("When a source do not 
receive any elements for the timeout time, " +
+   "it will be marked as 
temporarily idle. This allows downstream " +
+   "tasks to advance their 
watermarks without the need to wait for " +
+   "watermarks from this 
source while it is idle.");
+
+   // 

+   //  Sort Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SORT_DEFAULT_LIMIT 
=
+   key("table.exec.sort.default-limit")
+   .defaultValue(200)
+   .withDescription("Default limit when 
user don't set a limit after order by.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES =
+   key("table.exec.sort.max-num-file-handles")
+   .defaultValue(128)
+   .withDescription("The maximal fan-in 
for external merge sort. It limits the number of file handles per operator. " +
+   "If it is too small, 
may cause intermediate merging. But if it is too large, " +
+   "it will cause too many 
files opened at the same time, consume memory and lead to random reading.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_ASYNC_MERGE_ENABLED =
+   key("table.exec.sort.async-merge-enabled")
+   .defaultValue(true)
+   .withDescription("Whether to 
asynchronously merge sorted spill files.");
+
+   // 

+   //  Spill Options
+   // 

+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_ENABLED =
+   key("table.exec.spill-compression.enabled")
+   .defaultValue(true)
+   .withDescription("Whether to compress 
spilled data. " +
+   "(Now include sort and 
hash agg and hash join)");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_CODEC =
+   key("table.exec.spill-compression.codec")
+   .defaultValue("lz4")
+   .withDescription("Use that compression 
codec to compress spilled file. " +
+   "Now we only support 
lz4.");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_BLOCK_SIZE =

[GitHub] [flink] flinkbot commented on issue #9224: [FLINK-13266] [connectors / kafka] Fix race condition between transaction commit and produc…

2019-07-24 Thread GitBox
flinkbot commented on issue #9224: [FLINK-13266] [connectors / kafka] Fix race 
condition between transaction commit and produc…
URL: https://github.com/apache/flink/pull/9224#issuecomment-514888275
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307094870
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/OptimizerConfigOptions.java
 ##
 @@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table planner 
module.
+ */
+public class OptimizerConfigOptions {
+
+   // 

+   //  Optimizer Options
+   // 

+   public static final ConfigOption 
TABLE_OPTIMIZER_AGG_PHASE_STRATEGY =
+   key("table.optimizer.agg-phase-strategy")
+   .defaultValue("AUTO")
+   .withDescription("Strategy for 
aggregate phase. Only AUTO, TWO_PHASE or ONE_PHASE can be set.\n" +
+   "AUTO: No special 
enforcer for aggregate stage. Whether to choose two stage aggregate or one" +
+   " stage aggregate 
depends on cost. \n" +
+   "TWO_PHASE: Enforce to 
use two stage aggregate which has localAggregate and globalAggregate. " +
+   "Note that if aggregate 
call does not support optimize into two phase, we will still use one stage 
aggregate.\n" +
+   "ONE_PHASE: Enforce to 
use one stage aggregate which only has CompleteGlobalAggregate.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_BROADCAST_JOIN_THRESHOLD =
+   key("table.optimizer.join.broadcast-threshold")
+   .defaultValue(1024 * 1024L)
+   .withDescription("Configures the 
maximum size in bytes for a table that will be broadcast to all worker " +
+   "nodes when performing 
a join. By setting this value to -1 to disable broadcasting.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_ENABLED =
+   key("table.optimizer.distinct-agg.split.enabled")
+   .defaultValue(false)
+   .withDescription("Tells the optimizer 
whether to split distinct aggregation " +
+   "(e.g. COUNT(DISTINCT 
col), SUM(DISTINCT col)) into two level. " +
 
 Review comment:
two level =>  two phase ? keep concept consistent with 
`table.optimizer.agg-phase-strategy`


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307103832
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/ExecutionConfigOptions.java
 ##
 @@ -0,0 +1,202 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table module.
+ */
+public class ExecutionConfigOptions {
+
+   // 

+   //  Source Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SOURCE_IDLE_TIMEOUT 
=
+   key("table.exec.source.idle-timeout")
+   .defaultValue("-1 ms")
+   .withDescription("When a source do not 
receive any elements for the timeout time, " +
+   "it will be marked as 
temporarily idle. This allows downstream " +
+   "tasks to advance their 
watermarks without the need to wait for " +
+   "watermarks from this 
source while it is idle.");
+
+   // 

+   //  Sort Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SORT_DEFAULT_LIMIT 
=
+   key("table.exec.sort.default-limit")
+   .defaultValue(200)
+   .withDescription("Default limit when 
user don't set a limit after order by.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES =
+   key("table.exec.sort.max-num-file-handles")
+   .defaultValue(128)
+   .withDescription("The maximal fan-in 
for external merge sort. It limits the number of file handles per operator. " +
+   "If it is too small, 
may cause intermediate merging. But if it is too large, " +
+   "it will cause too many 
files opened at the same time, consume memory and lead to random reading.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_ASYNC_MERGE_ENABLED =
+   key("table.exec.sort.async-merge-enabled")
+   .defaultValue(true)
+   .withDescription("Whether to 
asynchronously merge sorted spill files.");
+
+   // 

+   //  Spill Options
+   // 

+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_ENABLED =
+   key("table.exec.spill-compression.enabled")
+   .defaultValue(true)
+   .withDescription("Whether to compress 
spilled data. " +
+   "(Now include sort and 
hash agg and hash join)");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_CODEC =
+   key("table.exec.spill-compression.codec")
+   .defaultValue("lz4")
+   .withDescription("Use that compression 
codec to compress spilled file. " +
+   "Now we only support 
lz4.");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_BLOCK_SIZE =

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307095384
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/OptimizerConfigOptions.java
 ##
 @@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table planner 
module.
+ */
+public class OptimizerConfigOptions {
+
+   // 

+   //  Optimizer Options
+   // 

+   public static final ConfigOption 
TABLE_OPTIMIZER_AGG_PHASE_STRATEGY =
+   key("table.optimizer.agg-phase-strategy")
+   .defaultValue("AUTO")
+   .withDescription("Strategy for 
aggregate phase. Only AUTO, TWO_PHASE or ONE_PHASE can be set.\n" +
+   "AUTO: No special 
enforcer for aggregate stage. Whether to choose two stage aggregate or one" +
+   " stage aggregate 
depends on cost. \n" +
+   "TWO_PHASE: Enforce to 
use two stage aggregate which has localAggregate and globalAggregate. " +
+   "Note that if aggregate 
call does not support optimize into two phase, we will still use one stage 
aggregate.\n" +
+   "ONE_PHASE: Enforce to 
use one stage aggregate which only has CompleteGlobalAggregate.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_BROADCAST_JOIN_THRESHOLD =
+   key("table.optimizer.join.broadcast-threshold")
+   .defaultValue(1024 * 1024L)
+   .withDescription("Configures the 
maximum size in bytes for a table that will be broadcast to all worker " +
+   "nodes when performing 
a join. By setting this value to -1 to disable broadcasting.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_ENABLED =
+   key("table.optimizer.distinct-agg.split.enabled")
+   .defaultValue(false)
+   .withDescription("Tells the optimizer 
whether to split distinct aggregation " +
+   "(e.g. COUNT(DISTINCT 
col), SUM(DISTINCT col)) into two level. " +
+   "The first aggregation 
is shuffled by an additional key which is calculated using " +
+   "the hashcode of 
distinct_key and number of buckets. This optimization is very useful " +
+   "when there is data 
skew in distinct aggregation and gives the ability to scale-up the job. " +
+   "Default is false.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_BUCKET_NUM =
+   key("table.optimizer.distinct-agg.split.bucket-num")
+   .defaultValue(1024)
+   .withDescription("Configure the number 
of buckets when splitting distinct aggregation. " +
+   "The number is used in 
the first level aggregation to calculate a bucket key " +
+   
"'hash_code(distinct_key) % BUCKET_NUM' which is used as an additional group 
key after splitting.");
+
+   public static final 

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307103221
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/ExecutionConfigOptions.java
 ##
 @@ -0,0 +1,202 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table module.
+ */
+public class ExecutionConfigOptions {
+
+   // 

+   //  Source Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SOURCE_IDLE_TIMEOUT 
=
+   key("table.exec.source.idle-timeout")
+   .defaultValue("-1 ms")
+   .withDescription("When a source do not 
receive any elements for the timeout time, " +
+   "it will be marked as 
temporarily idle. This allows downstream " +
+   "tasks to advance their 
watermarks without the need to wait for " +
+   "watermarks from this 
source while it is idle.");
+
+   // 

+   //  Sort Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SORT_DEFAULT_LIMIT 
=
+   key("table.exec.sort.default-limit")
+   .defaultValue(200)
+   .withDescription("Default limit when 
user don't set a limit after order by.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES =
+   key("table.exec.sort.max-num-file-handles")
+   .defaultValue(128)
+   .withDescription("The maximal fan-in 
for external merge sort. It limits the number of file handles per operator. " +
+   "If it is too small, 
may cause intermediate merging. But if it is too large, " +
+   "it will cause too many 
files opened at the same time, consume memory and lead to random reading.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_ASYNC_MERGE_ENABLED =
+   key("table.exec.sort.async-merge-enabled")
+   .defaultValue(true)
+   .withDescription("Whether to 
asynchronously merge sorted spill files.");
+
+   // 

+   //  Spill Options
+   // 

+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_ENABLED =
+   key("table.exec.spill-compression.enabled")
+   .defaultValue(true)
+   .withDescription("Whether to compress 
spilled data. " +
+   "(Now include sort and 
hash agg and hash join)");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_CODEC =
 
 Review comment:
   remove this option ?  no other choice now 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git 

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307095123
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/OptimizerConfigOptions.java
 ##
 @@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table planner 
module.
+ */
+public class OptimizerConfigOptions {
+
+   // 

+   //  Optimizer Options
+   // 

+   public static final ConfigOption 
TABLE_OPTIMIZER_AGG_PHASE_STRATEGY =
+   key("table.optimizer.agg-phase-strategy")
+   .defaultValue("AUTO")
+   .withDescription("Strategy for 
aggregate phase. Only AUTO, TWO_PHASE or ONE_PHASE can be set.\n" +
+   "AUTO: No special 
enforcer for aggregate stage. Whether to choose two stage aggregate or one" +
+   " stage aggregate 
depends on cost. \n" +
+   "TWO_PHASE: Enforce to 
use two stage aggregate which has localAggregate and globalAggregate. " +
+   "Note that if aggregate 
call does not support optimize into two phase, we will still use one stage 
aggregate.\n" +
+   "ONE_PHASE: Enforce to 
use one stage aggregate which only has CompleteGlobalAggregate.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_BROADCAST_JOIN_THRESHOLD =
+   key("table.optimizer.join.broadcast-threshold")
+   .defaultValue(1024 * 1024L)
+   .withDescription("Configures the 
maximum size in bytes for a table that will be broadcast to all worker " +
+   "nodes when performing 
a join. By setting this value to -1 to disable broadcasting.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_ENABLED =
+   key("table.optimizer.distinct-agg.split.enabled")
+   .defaultValue(false)
+   .withDescription("Tells the optimizer 
whether to split distinct aggregation " +
+   "(e.g. COUNT(DISTINCT 
col), SUM(DISTINCT col)) into two level. " +
+   "The first aggregation 
is shuffled by an additional key which is calculated using " +
+   "the hashcode of 
distinct_key and number of buckets. This optimization is very useful " +
+   "when there is data 
skew in distinct aggregation and gives the ability to scale-up the job. " +
+   "Default is false.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_BUCKET_NUM =
+   key("table.optimizer.distinct-agg.split.bucket-num")
+   .defaultValue(1024)
+   .withDescription("Configure the number 
of buckets when splitting distinct aggregation. " +
+   "The number is used in 
the first level aggregation to calculate a bucket key " +
 
 Review comment:
   first level => first phase ?


This is an automated message from the Apache Git Service.
To respond to the 

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307103492
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/ExecutionConfigOptions.java
 ##
 @@ -0,0 +1,202 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table module.
+ */
+public class ExecutionConfigOptions {
+
+   // 

+   //  Source Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SOURCE_IDLE_TIMEOUT 
=
+   key("table.exec.source.idle-timeout")
+   .defaultValue("-1 ms")
+   .withDescription("When a source do not 
receive any elements for the timeout time, " +
+   "it will be marked as 
temporarily idle. This allows downstream " +
+   "tasks to advance their 
watermarks without the need to wait for " +
+   "watermarks from this 
source while it is idle.");
+
+   // 

+   //  Sort Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SORT_DEFAULT_LIMIT 
=
+   key("table.exec.sort.default-limit")
+   .defaultValue(200)
+   .withDescription("Default limit when 
user don't set a limit after order by.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES =
+   key("table.exec.sort.max-num-file-handles")
+   .defaultValue(128)
+   .withDescription("The maximal fan-in 
for external merge sort. It limits the number of file handles per operator. " +
+   "If it is too small, 
may cause intermediate merging. But if it is too large, " +
+   "it will cause too many 
files opened at the same time, consume memory and lead to random reading.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_ASYNC_MERGE_ENABLED =
+   key("table.exec.sort.async-merge-enabled")
+   .defaultValue(true)
+   .withDescription("Whether to 
asynchronously merge sorted spill files.");
+
+   // 

+   //  Spill Options
+   // 

+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_ENABLED =
+   key("table.exec.spill-compression.enabled")
+   .defaultValue(true)
+   .withDescription("Whether to compress 
spilled data. " +
+   "(Now include sort and 
hash agg and hash join)");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_CODEC =
+   key("table.exec.spill-compression.codec")
+   .defaultValue("lz4")
+   .withDescription("Use that compression 
codec to compress spilled file. " +
+   "Now we only support 
lz4.");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_BLOCK_SIZE =

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307095535
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/OptimizerConfigOptions.java
 ##
 @@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table planner 
module.
+ */
+public class OptimizerConfigOptions {
+
+   // 

+   //  Optimizer Options
+   // 

+   public static final ConfigOption 
TABLE_OPTIMIZER_AGG_PHASE_STRATEGY =
+   key("table.optimizer.agg-phase-strategy")
+   .defaultValue("AUTO")
+   .withDescription("Strategy for 
aggregate phase. Only AUTO, TWO_PHASE or ONE_PHASE can be set.\n" +
+   "AUTO: No special 
enforcer for aggregate stage. Whether to choose two stage aggregate or one" +
+   " stage aggregate 
depends on cost. \n" +
+   "TWO_PHASE: Enforce to 
use two stage aggregate which has localAggregate and globalAggregate. " +
+   "Note that if aggregate 
call does not support optimize into two phase, we will still use one stage 
aggregate.\n" +
+   "ONE_PHASE: Enforce to 
use one stage aggregate which only has CompleteGlobalAggregate.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_BROADCAST_JOIN_THRESHOLD =
+   key("table.optimizer.join.broadcast-threshold")
+   .defaultValue(1024 * 1024L)
+   .withDescription("Configures the 
maximum size in bytes for a table that will be broadcast to all worker " +
+   "nodes when performing 
a join. By setting this value to -1 to disable broadcasting.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_ENABLED =
+   key("table.optimizer.distinct-agg.split.enabled")
+   .defaultValue(false)
+   .withDescription("Tells the optimizer 
whether to split distinct aggregation " +
+   "(e.g. COUNT(DISTINCT 
col), SUM(DISTINCT col)) into two level. " +
+   "The first aggregation 
is shuffled by an additional key which is calculated using " +
+   "the hashcode of 
distinct_key and number of buckets. This optimization is very useful " +
+   "when there is data 
skew in distinct aggregation and gives the ability to scale-up the job. " +
+   "Default is false.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_BUCKET_NUM =
+   key("table.optimizer.distinct-agg.split.bucket-num")
+   .defaultValue(1024)
+   .withDescription("Configure the number 
of buckets when splitting distinct aggregation. " +
+   "The number is used in 
the first level aggregation to calculate a bucket key " +
+   
"'hash_code(distinct_key) % BUCKET_NUM' which is used as an additional group 
key after splitting.");
+
+   public static final 

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307095307
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/OptimizerConfigOptions.java
 ##
 @@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table planner 
module.
+ */
+public class OptimizerConfigOptions {
+
+   // 

+   //  Optimizer Options
+   // 

+   public static final ConfigOption 
TABLE_OPTIMIZER_AGG_PHASE_STRATEGY =
+   key("table.optimizer.agg-phase-strategy")
+   .defaultValue("AUTO")
+   .withDescription("Strategy for 
aggregate phase. Only AUTO, TWO_PHASE or ONE_PHASE can be set.\n" +
+   "AUTO: No special 
enforcer for aggregate stage. Whether to choose two stage aggregate or one" +
+   " stage aggregate 
depends on cost. \n" +
+   "TWO_PHASE: Enforce to 
use two stage aggregate which has localAggregate and globalAggregate. " +
+   "Note that if aggregate 
call does not support optimize into two phase, we will still use one stage 
aggregate.\n" +
+   "ONE_PHASE: Enforce to 
use one stage aggregate which only has CompleteGlobalAggregate.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_BROADCAST_JOIN_THRESHOLD =
+   key("table.optimizer.join.broadcast-threshold")
+   .defaultValue(1024 * 1024L)
+   .withDescription("Configures the 
maximum size in bytes for a table that will be broadcast to all worker " +
+   "nodes when performing 
a join. By setting this value to -1 to disable broadcasting.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_ENABLED =
+   key("table.optimizer.distinct-agg.split.enabled")
+   .defaultValue(false)
+   .withDescription("Tells the optimizer 
whether to split distinct aggregation " +
+   "(e.g. COUNT(DISTINCT 
col), SUM(DISTINCT col)) into two level. " +
+   "The first aggregation 
is shuffled by an additional key which is calculated using " +
+   "the hashcode of 
distinct_key and number of buckets. This optimization is very useful " +
+   "when there is data 
skew in distinct aggregation and gives the ability to scale-up the job. " +
+   "Default is false.");
+
+   public static final ConfigOption 
TABLE_OPTIMIZER_DISTINCT_AGG_SPLIT_BUCKET_NUM =
+   key("table.optimizer.distinct-agg.split.bucket-num")
+   .defaultValue(1024)
+   .withDescription("Configure the number 
of buckets when splitting distinct aggregation. " +
+   "The number is used in 
the first level aggregation to calculate a bucket key " +
+   
"'hash_code(distinct_key) % BUCKET_NUM' which is used as an additional group 
key after splitting.");
+
+   public static final 

[GitHub] [flink] godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
godfreyhe commented on a change in pull request #9203: [FLINK-13375][table-api] 
Improve config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#discussion_r307106953
 
 

 ##
 File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/config/ExecutionConfigOptions.java
 ##
 @@ -0,0 +1,202 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.api.config;
+
+import org.apache.flink.configuration.ConfigOption;
+
+import static org.apache.flink.configuration.ConfigOptions.key;
+
+/**
+ * This class holds configuration constants used by Flink's table module.
+ */
+public class ExecutionConfigOptions {
+
+   // 

+   //  Source Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SOURCE_IDLE_TIMEOUT 
=
+   key("table.exec.source.idle-timeout")
+   .defaultValue("-1 ms")
+   .withDescription("When a source do not 
receive any elements for the timeout time, " +
+   "it will be marked as 
temporarily idle. This allows downstream " +
+   "tasks to advance their 
watermarks without the need to wait for " +
+   "watermarks from this 
source while it is idle.");
+
+   // 

+   //  Sort Options
+   // 

+   public static final ConfigOption TABLE_EXEC_SORT_DEFAULT_LIMIT 
=
+   key("table.exec.sort.default-limit")
+   .defaultValue(200)
+   .withDescription("Default limit when 
user don't set a limit after order by.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_MAX_NUM_FILE_HANDLES =
+   key("table.exec.sort.max-num-file-handles")
+   .defaultValue(128)
+   .withDescription("The maximal fan-in 
for external merge sort. It limits the number of file handles per operator. " +
+   "If it is too small, 
may cause intermediate merging. But if it is too large, " +
+   "it will cause too many 
files opened at the same time, consume memory and lead to random reading.");
+
+   public static final ConfigOption 
TABLE_EXEC_SORT_ASYNC_MERGE_ENABLED =
+   key("table.exec.sort.async-merge-enabled")
+   .defaultValue(true)
+   .withDescription("Whether to 
asynchronously merge sorted spill files.");
+
+   // 

+   //  Spill Options
+   // 

+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_ENABLED =
+   key("table.exec.spill-compression.enabled")
+   .defaultValue(true)
+   .withDescription("Whether to compress 
spilled data. " +
+   "(Now include sort and 
hash agg and hash join)");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_CODEC =
+   key("table.exec.spill-compression.codec")
+   .defaultValue("lz4")
+   .withDescription("Use that compression 
codec to compress spilled file. " +
+   "Now we only support 
lz4.");
+
+   public static final ConfigOption 
TABLE_EXEC_SPILL_COMPRESSION_BLOCK_SIZE =

[GitHub] [flink] becketqin opened a new pull request #9224: [FLINK-13266] [connectors / kafka] Fix race condition between transaction commit and produc…

2019-07-24 Thread GitBox
becketqin opened a new pull request #9224: [FLINK-13266] [connectors / kafka] 
Fix race condition between transaction commit and produc…
URL: https://github.com/apache/flink/pull/9224
 
 
   …er closure.
   
   ## What is the purpose of the change
   This patch fixes a race condition between the checkpointing thread and main 
thread. The sequence causing the deadlock is the following:
   1. In `FlinkKafkaProducer`, the main thread encounters a problem and closes 
all the producer to start failover.
   2. The previous checkpoint has completed, so the checkpointing thread grabs 
the checkpoint lock and tries to commit the transaction on the producer that 
has been closed in step 1. This commit will never succeed due to 
[KAFKA-6635](https://issues.apache.org/jira/browse/KAFKA-6635). So the 
checkpoint thread blocks forever.
   3. In `StreamTask`, the main thread will eventually try to release all the 
record writer. To do that, it attempts to grab the checkpoint lock which is 
hold by checkpoint thread in step 2 and will never be released. So the main 
thread also blocks forever.
   
   KAFKA-6635 has been fixed in Kafka 2.3.0. But Flink 1.9 does not rely on 
that yet, and we also support Kafka 0.11. So we are just going to fix on the 
Flink side first. The solution is to make sure that in `FlinkKafkaProducer` any 
operation relying on the underlying sender thread to finish throws an exception 
if the producer is closed.
   
   This patch also fixes a minor issue of duplicated static inner class name in 
`KafkaConsumerTestBase`.
   
   ## Brief change log
   - Make `FlinkKafkaProducer` and `FlinkKafkaInternalProducer` thread safe.
   - Fix static inner class name collision in `KafkaConsumerTestBase`.
   
   
   ## Verifying this change
   
   This change added tests and can be verified as follows:
   
   - Added test to ensure exception will be thrown if a blocking method is 
called on `FlinkKafkaProducer` and `FlinkKafkaInternalProducer` after the 
producer is closed.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-24 Thread GitBox
lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test 
should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-514888066
 
 
   Travis build on my own repo finished: 
https://travis-ci.org/lirui-apache/flink/builds/563356262
   Hive connector tests passed and I don't think other failures are related to 
this PR


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-13415) Document how to use hive connector in scala shell

2019-07-24 Thread Jeff Zhang (JIRA)
Jeff Zhang created FLINK-13415:
--

 Summary: Document how to use hive connector in scala shell
 Key: FLINK-13415
 URL: https://issues.apache.org/jira/browse/FLINK-13415
 Project: Flink
  Issue Type: Improvement
  Components: Documentation
Affects Versions: 1.9.0
Reporter: Jeff Zhang






--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9175: [FLINK-12038] [test] fix YARNITCase random fail

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9175: [FLINK-12038] [test] fix YARNITCase 
random fail
URL: https://github.com/apache/flink/pull/9175#issuecomment-513163926
 
 
   ## CI report:
   
   * f236f1934804c0b9b2d5df4564fad5f2c5ed9821 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766106)
   * 0d91c6c63c1e07526a3f9441f0ec5c439fb48dec : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120124248)
   * f6fa5e97c05927fbccf02f586f7c3b26303e6efd : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120426579)
   * 9f3c1c71b93b1813c85bb90bf4e2ef4b69e7 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120547384)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13410) Csv input format does not support LocalDate

2019-07-24 Thread Jingsong Lee (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892389#comment-16892389
 ] 

Jingsong Lee commented on FLINK-13410:
--

Hi [~TsReaper] blink support connectors related issues, please create them 
under https://issues.apache.org/jira/browse/FLINK-13285

> Csv input format does not support LocalDate
> ---
>
> Key: FLINK-13410
> URL: https://issues.apache.org/jira/browse/FLINK-13410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>
> Csv input format is lacking parsers for LocalDate, LocalTime, etc. As 
> DataTypes.DATE now defaults to LocalDate, we should add these parsers for 
> user experience.
> A temporal workaround for the users is that, users can call 
> DataTypes.Timestamp().bridgeTo(java.sql.Timestamp) to use the old 
> SqlTimestamp converter.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #7262: [FLINK-10478] Kafka Producer wrongly formats % for transaction ID

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #7262: [FLINK-10478] Kafka Producer wrongly 
formats % for transaction ID
URL: https://github.com/apache/flink/pull/7262#issuecomment-510752916
 
 
   ## CI report:
   
   * 15f05f4c5791d9d42610099324e59057d26bd3ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/118882595)
   * 2fc680ca11665e05fe0a66f3a92ba5e5d2d734cc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541120)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9211: [FLINK-13393][FLINK-13391][table-planner-blink] Fix source conversion and source return type

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9211: 
[FLINK-13393][FLINK-13391][table-planner-blink] Fix source conversion and 
source return type
URL: https://github.com/apache/flink/pull/9211#issuecomment-514475275
 
 
   ## CI report:
   
   * b465c69b21a3ac810bc34e4f95dc0b3a3d93281c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120311139)
   * 645face305b998b122ee5fb2d731926179885432 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120431665)
   * 7dfc56376b2c4ec37fb38c91851e20e76128ad7c : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120547055)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] shuai-xu commented on a change in pull request #9175: [FLINK-12038] [test] fix YARNITCase random fail

2019-07-24 Thread GitBox
shuai-xu commented on a change in pull request #9175: [FLINK-12038] [test] fix 
YARNITCase random fail
URL: https://github.com/apache/flink/pull/9175#discussion_r307105775
 
 

 ##
 File path: 
flink-yarn-tests/src/test/java/org/apache/flink/yarn/YarnTestBase.java
 ##
 @@ -573,6 +576,23 @@ public static int getRunningContainers() {
return count;
}
 
+   void waitUntilApplicationFinished(ApplicationId applicationId, Duration 
timeout) throws Exception {
+   Deadline deadline = Deadline.now().plus(timeout);
+   while (true) {
 
 Review comment:
   This may lead to unnecessary `sleep`, I change it to `while (state != 
YarnApplicationState.FINISHED) {}`


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] shuai-xu commented on a change in pull request #9175: [FLINK-12038] [test] fix YARNITCase random fail

2019-07-24 Thread GitBox
shuai-xu commented on a change in pull request #9175: [FLINK-12038] [test] fix 
YARNITCase random fail
URL: https://github.com/apache/flink/pull/9175#discussion_r307105587
 
 

 ##
 File path: 
flink-yarn-tests/src/test/java/org/apache/flink/yarn/YarnTestBase.java
 ##
 @@ -573,6 +576,23 @@ public static int getRunningContainers() {
return count;
}
 
+   void waitUntilApplicationFinished(ApplicationId applicationId, Duration 
timeout) throws Exception {
+   Deadline deadline = Deadline.now().plus(timeout);
+   while (true) {
+   YarnApplicationState state = 
yarnClient.getApplicationReport(applicationId).getYarnApplicationState();
+   if (state == YarnApplicationState.FINISHED) {
+   break;
+   } else if (state == YarnApplicationState.FAILED || 
state == YarnApplicationState.KILLED) {
+   Assert.fail("Application became FAILED or 
KILLED while expecting FINISHED");
+   } else {
+   sleep(sleepIntervalInMS);
+   }
+   if (deadline.isOverdue()) {
+   Assert.fail("Application didn't finish before 
timeout");
 
 Review comment:
   I put the `waitUntilApplicationFinished` into YarnITCase as it is not so 
universal for other case now.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9195: [FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert table sink

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9195: [FLINK-13289][table-planner-blink] 
Blink-planner should setKeyFields to upsert table sink
URL: https://github.com/apache/flink/pull/9195#issuecomment-513690205
 
 
   ## CI report:
   
   * 5a275f751d5b7e13cbf4ce2ecacaf923bc8ccaa8 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119973212)
   * e959fd39d8a2d34b92591d2cd31ec405625c85b4 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120119338)
   * 094123574c35c48ce7c93c1e47cd35a81818ad40 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120429194)
   * 4c5fe1ac569d3f4a8ae11f8eeaf4d7090baf425c : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120546407)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] JingsongLi commented on issue #9195: [FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert table sink

2019-07-24 Thread GitBox
JingsongLi commented on issue #9195: [FLINK-13289][table-planner-blink] 
Blink-planner should setKeyFields to upsert table sink
URL: https://github.com/apache/flink/pull/9195#issuecomment-514883460
 
 
   @wuchong Please take a look again.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9222: [FLINK-13409][python] Supported java 
UDFs in python API.
URL: https://github.com/apache/flink/pull/9222#issuecomment-514866924
 
 
   ## CI report:
   
   * 69b88611a023533b59dee72fe56332d77ee93884 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541108)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] JingsongLi commented on a change in pull request #9195: [FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert table sink

2019-07-24 Thread GitBox
JingsongLi commented on a change in pull request #9195: 
[FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert 
table sink
URL: https://github.com/apache/flink/pull/9195#discussion_r307103615
 
 

 ##
 File path: 
flink-table/flink-table-planner-blink/src/test/scala/org/apache/flink/table/runtime/stream/table/TableSinkITCase.scala
 ##
 @@ -0,0 +1,701 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.runtime.stream.table
+
+import org.apache.flink.api.common.functions.MapFunction
+import org.apache.flink.api.common.typeinfo.TypeInformation
+import org.apache.flink.api.java.tuple.{Tuple2 => JTuple2}
+import org.apache.flink.api.java.typeutils.RowTypeInfo
+import org.apache.flink.api.scala._
+import org.apache.flink.streaming.api.TimeCharacteristic
+import org.apache.flink.streaming.api.datastream.{DataStream, DataStreamSink}
+import org.apache.flink.streaming.api.functions.sink.SinkFunction
+import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment
+import org.apache.flink.table.api.scala._
+import org.apache.flink.table.api.{TableException, Tumble, Types}
+import org.apache.flink.table.planner.runtime.utils.TestData.{smallTupleData3, 
tupleData3, tupleData5}
+import org.apache.flink.table.planner.utils.MemoryTableSourceSinkUtil
+import org.apache.flink.table.sinks._
+import org.apache.flink.test.util.{AbstractTestBase, TestBaseUtils}
+import org.apache.flink.types.Row
+
+import org.junit.Assert._
+import org.junit.Test
+
+import java.io.File
+import java.lang.{Boolean => JBool}
+
+import scala.collection.JavaConverters._
+import scala.collection.mutable
+
+class TableSinkITCase extends AbstractTestBase {
+
+  @Test
+  def testInsertIntoRegisteredTableSink(): Unit = {
+val env = StreamExecutionEnvironment.getExecutionEnvironment
+env.getConfig.enableObjectReuse()
+env.setStreamTimeCharacteristic(TimeCharacteristic.EventTime)
+
+val tEnv = StreamTableEnvironment.create(env)
+MemoryTableSourceSinkUtil.clear()
+
+val input = env.fromCollection(tupleData3)
+  .assignAscendingTimestamps(r => r._2)
+val fieldNames = Array("d", "e", "t")
+val fieldTypes: Array[TypeInformation[_]] = Array(Types.STRING, 
Types.SQL_TIMESTAMP, Types.LONG)
+val sink = new MemoryTableSourceSinkUtil.UnsafeMemoryAppendTableSink
+tEnv.registerTableSink("targetTable", sink.configure(fieldNames, 
fieldTypes))
+
+input.toTable(tEnv, 'a, 'b, 'c, 't.rowtime)
+  .where('a < 3 || 'a > 19)
+  .select('c, 't, 'b)
+  .insertInto("targetTable")
+env.execute()
+
+val expected = Seq(
+  "Hi,1970-01-01 00:00:00.001,1",
+  "Hello,1970-01-01 00:00:00.002,2",
+  "Comment#14,1970-01-01 00:00:00.006,6",
+  "Comment#15,1970-01-01 00:00:00.006,6").mkString("\n")
+
+
TestBaseUtils.compareResultAsText(MemoryTableSourceSinkUtil.tableData.asJava, 
expected)
+  }
+
+  @Test
+  def testStreamTableSink(): Unit = {
+
+val tmpFile = File.createTempFile("flink-table-sink-test", ".tmp")
+tmpFile.deleteOnExit()
+val path = tmpFile.toURI.toString
+
+val env = StreamExecutionEnvironment.getExecutionEnvironment
+env.getConfig.enableObjectReuse()
+env.setStreamTimeCharacteristic(TimeCharacteristic.EventTime)
+
+val tEnv = StreamTableEnvironment.create(env)
+env.setParallelism(4)
+
+tEnv.registerTableSink(
+  "csvSink",
+  new CsvTableSink(path).configure(
+Array[String]("c", "b"),
+Array[TypeInformation[_]](Types.STRING, Types.SQL_TIMESTAMP)))
+
+val input = env.fromCollection(tupleData3)
+  .assignAscendingTimestamps(_._2)
+  .map(x => x).setParallelism(4) // increase DOP to 4
+
+input.toTable(tEnv, 'a, 'b.rowtime, 'c)
+  .where('a < 5 || 'a > 17)
+  .select('c, 'b)
+  .insertInto("csvSink")
+
+env.execute()
+
+val expected = Seq(
+  "Hi,1970-01-01 00:00:00.001",
+  "Hello,1970-01-01 00:00:00.002",
+  "Hello world,1970-01-01 00:00:00.002",
+  "Hello world, how are you?,1970-01-01 00:00:00.003",
+  "Comment#12,1970-01-01 00:00:00.006",
+  "Comment#13,1970-01-01 00:00:00.006",
+  "Comment#14,1970-01-01 00:00:00.006",
+

[jira] [Commented] (FLINK-13414) Add support for Scala 2.13

2019-07-24 Thread Chaoran Yu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13414?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892384#comment-16892384
 ] 

Chaoran Yu commented on FLINK-13414:


Is any committer or someone in the community working on it? It will be great to 
support Scala 2.13 in Flink.

> Add support for Scala 2.13
> --
>
> Key: FLINK-13414
> URL: https://issues.apache.org/jira/browse/FLINK-13414
> Project: Flink
>  Issue Type: Task
>  Components: API / Scala
>Reporter: Chaoran Yu
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Created] (FLINK-13414) Add support for Scala 2.13

2019-07-24 Thread Chaoran Yu (JIRA)
Chaoran Yu created FLINK-13414:
--

 Summary: Add support for Scala 2.13
 Key: FLINK-13414
 URL: https://issues.apache.org/jira/browse/FLINK-13414
 Project: Flink
  Issue Type: Task
  Components: API / Scala
Reporter: Chaoran Yu






--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9141: [FLINK-12249][table] Fix type equivalence check problems for Window Aggregates

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9141: [FLINK-12249][table] Fix type 
equivalence check problems for Window Aggregates
URL: https://github.com/apache/flink/pull/9141#issuecomment-512181311
 
 
   ## CI report:
   
   * 1af43966364d445984524fd3bf51a0f3b8e75cbe : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119443400)
   * d7823627d50063f634d1ad6e43f470947a54daed : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119585913)
   * caf138832a23ecc7a3f72ffef760e5a3b070e0f1 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766257)
   * c6cd8ac0437314bf69074625d1a81dc72c664540 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120390913)
   * 5e29a7430750ece078ecf435938dde4e1a21366c : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120540112)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13410) Csv input format does not support LocalDate

2019-07-24 Thread Caizhi Weng (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892381#comment-16892381
 ] 

Caizhi Weng commented on FLINK-13410:
-

They are sort of related because they are the same type of problem. But the 
solutions will be different and separate.

> Csv input format does not support LocalDate
> ---
>
> Key: FLINK-13410
> URL: https://issues.apache.org/jira/browse/FLINK-13410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>
> Csv input format is lacking parsers for LocalDate, LocalTime, etc. As 
> DataTypes.DATE now defaults to LocalDate, we should add these parsers for 
> user experience.
> A temporal workaround for the users is that, users can call 
> DataTypes.Timestamp().bridgeTo(java.sql.Timestamp) to use the old 
> SqlTimestamp converter.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Created] (FLINK-13413) Put flink-table jar under lib folder

2019-07-24 Thread Jeff Zhang (JIRA)
Jeff Zhang created FLINK-13413:
--

 Summary: Put flink-table jar under lib folder
 Key: FLINK-13413
 URL: https://issues.apache.org/jira/browse/FLINK-13413
 Project: Flink
  Issue Type: Improvement
  Components: Release System
Affects Versions: 1.9.0
Reporter: Jeff Zhang


Now flink-table jar is in opt folder. Since we plan to make flink-table api as 
the first citizen of flink, we should put it under lib folder, otherwise it is 
not convenient to use some tools that use flink table api, such as flink scala 
shell, sql client and etc. 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13338) Sql conformance is hard to config in TableConfig

2019-07-24 Thread Kurt Young (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13338?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892374#comment-16892374
 ] 

Kurt Young commented on FLINK-13338:


Could we just set default conformance to HIVE? Currently it seems unnecessary 
to config conformance. 

> Sql conformance is hard to config in TableConfig
> 
>
> Key: FLINK-13338
> URL: https://issues.apache.org/jira/browse/FLINK-13338
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Danny Chan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Now the TableConfig has only interface to config the SqlParser config which 
> is very broad and hard to use for user, we should at least supply an 
> interface to config the sql conformance.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot commented on issue #9223: [FLINK-13352][hive] Using hive connector with hive-1.2.1 needs libfb3…

2019-07-24 Thread GitBox
flinkbot commented on issue #9223: [FLINK-13352][hive] Using hive connector 
with hive-1.2.1 needs libfb3…
URL: https://github.com/apache/flink/pull/9223#issuecomment-514873847
 
 
   ## CI report:
   
   * 3ac1a7e5322324b5ce2fd0aaf14cfbecc4793670 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120543108)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-13412) Test fail on ARM

2019-07-24 Thread wangxiyuan (JIRA)
wangxiyuan created FLINK-13412:
--

 Summary: Test fail on ARM
 Key: FLINK-13412
 URL: https://issues.apache.org/jira/browse/FLINK-13412
 Project: Flink
  Issue Type: Bug
  Components: Tests
Affects Versions: 1.9.0
Reporter: wangxiyuan
 Fix For: 1.9.0


We hit some test failures when running `mvn verify` for Flink on ARM arch.

--
1. org.apache.flink.util.MemoryArchitectureTest

[ERROR] 
testArchitectureNotUnknown(org.apache.flink.util.MemoryArchitectureTest) Time 
elapsed: 0.024 s <<< FAILURE!
java.lang.AssertionError: Values should be different. Actual: UNKNOWN
at 
org.apache.flink.util.MemoryArchitectureTest.testArchitectureNotUnknown(MemoryArchitectureTest.java:34)

 

2. org.apache.flink.table.expressions.SqlExpressionTest

testArithmeticFunctions org.junit.ComparisonFailure: Wrong result for: 
[LOG(3,27)] optimized to: [3.0004440E0] expected:<3.0[]> but 
was:<3.0[004]>

--

For the first one, I think we should add `aarch64` to the knonwn arch list as 
well.

 

For the second one, the logarithms behavior is a little different between arm 
and x86 .For example log(3,27) or log(7,343), the result is 3.0 on x86 but 
3.0004 on arm,  can we remove the test log(3,27), or change to 
something like log(2,8) ?



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] lirui-apache commented on issue #9223: [FLINK-13352][hive] Using hive connector with hive-1.2.1 needs libfb3…

2019-07-24 Thread GitBox
lirui-apache commented on issue #9223: [FLINK-13352][hive] Using hive connector 
with hive-1.2.1 needs libfb3…
URL: https://github.com/apache/flink/pull/9223#issuecomment-514873082
 
 
   cc @xuefuz @zjuwangg 
   I also find the hierarchy about the HiveCatalog is a little strange -- we 
have 2 `HiveCatalog` headings. Not sure if we can use this PR to fix that as 
well.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13410) Csv input format does not support LocalDate

2019-07-24 Thread Kurt Young (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892368#comment-16892368
 ] 

Kurt Young commented on FLINK-13410:


Is this related to https://issues.apache.org/jira/browse/FLINK-13283 ?

> Csv input format does not support LocalDate
> ---
>
> Key: FLINK-13410
> URL: https://issues.apache.org/jira/browse/FLINK-13410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>
> Csv input format is lacking parsers for LocalDate, LocalTime, etc. As 
> DataTypes.DATE now defaults to LocalDate, we should add these parsers for 
> user experience.
> A temporal workaround for the users is that, users can call 
> DataTypes.Timestamp().bridgeTo(java.sql.Timestamp) to use the old 
> SqlTimestamp converter.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot commented on issue #9223: [FLINK-13352][hive] Using hive connector with hive-1.2.1 needs libfb3…

2019-07-24 Thread GitBox
flinkbot commented on issue #9223: [FLINK-13352][hive] Using hive connector 
with hive-1.2.1 needs libfb3…
URL: https://github.com/apache/flink/pull/9223#issuecomment-514872882
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13352) Using hive connector with hive-1.2.1 needs libfb303 jar

2019-07-24 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13352?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-13352:
---
Labels: pull-request-available  (was: )

> Using hive connector with hive-1.2.1 needs libfb303 jar
> ---
>
> Key: FLINK-13352
> URL: https://issues.apache.org/jira/browse/FLINK-13352
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive, Documentation
>Reporter: Rui Li
>Priority: Major
>  Labels: pull-request-available
>
> Should mention that libfb303 jar is needed in {{catalog.md}}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] lirui-apache opened a new pull request #9223: [FLINK-13352][hive] Using hive connector with hive-1.2.1 needs libfb3…

2019-07-24 Thread GitBox
lirui-apache opened a new pull request #9223: [FLINK-13352][hive] Using hive 
connector with hive-1.2.1 needs libfb3…
URL: https://github.com/apache/flink/pull/9223
 
 
   …03 jar
   
   
   
   ## What is the purpose of the change
   
   To document that libfb303 is needed when using hive connector  with 
hive-1.2.1.
   
   ## Brief change log
   
 - Updated the dependencies part for hive-1.2.1.
   
   
   ## Verifying this change
   
   Trivial doc change
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? NA
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13401) Flink SQL sqlUpdate sql where complex conditional expressions do not take effect

2019-07-24 Thread Kurt Young (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13401?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892364#comment-16892364
 ] 

Kurt Young commented on FLINK-13401:


Could you try 1.9.0 and see if the result still not match with mysql?

> Flink SQL sqlUpdate sql where complex conditional expressions do not take 
> effect
> 
>
> Key: FLINK-13401
> URL: https://issues.apache.org/jira/browse/FLINK-13401
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Affects Versions: 1.6.3, 1.7.2
>Reporter: pingle wang
>Priority: Major
> Fix For: 1.9.0
>
> Attachments: image-2019-07-24-16-40-36-070.png, 
> image-2019-07-24-16-45-20-682.png, image-2019-07-24-17-00-36-079.png
>
>
> flink execute sql :
>  
> {code:java}
> select
> count(ad_action) as cnt
> from
> ad_sdk_android
> where
> ad_position_id = '100055'
> and ad_action = 'click'
> and (event_type IS NULL OR event_type <> '2'AND event_type <> '10')
> {code}
> the result is 20, but in mysql and hive, this result is 17.
> !image-2019-07-24-17-00-36-079.png!
>  
> The key condition is :
>  
> {code:java}
> and (event_type IS NULL OR event_type <> '2'AND event_type <> '10'){code}
> and event_type is varchar, flink is string data type.
>  
> mysql has the key condition sql result is:
> !image-2019-07-24-16-40-36-070.png!
> mysql no the key condition sql result is:
> !image-2019-07-24-16-45-20-682.png!
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] JingsongLi commented on a change in pull request #9195: [FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert table sink

2019-07-24 Thread GitBox
JingsongLi commented on a change in pull request #9195: 
[FLINK-13289][table-planner-blink] Blink-planner should setKeyFields to upsert 
table sink
URL: https://github.com/apache/flink/pull/9195#discussion_r307093662
 
 

 ##
 File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamExecSink.scala
 ##
 @@ -97,6 +97,22 @@ class StreamExecSink[T](
 // check for append only table
 val isAppendOnlyTable = UpdatingPlanChecker.isAppendOnly(this)
 upsertSink.setIsAppendOnly(isAppendOnlyTable)
+
+// extract unique key fields
+// Now we pick shortest one to sink
+// TODO UpsertStreamTableSink setKeyFields interface should be 
Array[Array[String]]
+val tableKeys: Option[Array[String]] =
+  UpdatingPlanChecker.getUniqueKeys(getInput, 
planner).sortBy(_.length).headOption
 
 Review comment:
   I think we can return `Option[Array[Array[String]]]`, let `sortBy and 
headOption` still remain in caller, keep the TODO with 
`UpsertStreamTableSink.setKeyFields`


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-13411) SqlClient swallow the error when fails to submit job

2019-07-24 Thread Jeff Zhang (JIRA)
Jeff Zhang created FLINK-13411:
--

 Summary: SqlClient swallow the error when fails to submit job
 Key: FLINK-13411
 URL: https://issues.apache.org/jira/browse/FLINK-13411
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / Client
Affects Versions: 1.9.0
Reporter: Jeff Zhang


I start sql-client without starting flink cluster. So when I submit sql in 
sql-client, it should throw exception to tell end-user that it fails to submit 
job, but what I see is that the sql-client display nothing. This is not 
user-friendly. 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9181: [FLINK-13210][hive] Hive connector 
test should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-513213135
 
 
   ## CI report:
   
   * d9311024212f04cca28d1105ab51512c4de4aedd : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119788315)
   * 8a12e7fc07dff371615365cfcbea038a02c232df : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119875298)
   * c4af2954d620a1fc0e8fb84443a6e09a7f258e66 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120364607)
   * a2d681a2a698cc1ff6053afcfd4cdafe392626dd : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541820)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] dianfu commented on issue #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
dianfu commented on issue #9222: [FLINK-13409][python] Supported java UDFs in 
python API.
URL: https://github.com/apache/flink/pull/9222#issuecomment-514869142
 
 
   @WeiZhong94 Thanks a lot for the PR. It would be better to also update the 
docs.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-24 Thread GitBox
lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test 
should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-514867852
 
 
   @wuchong it's #9216. I just rebased the PR


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
flinkbot commented on issue #9222: [FLINK-13409][python] Supported java UDFs in 
python API.
URL: https://github.com/apache/flink/pull/9222#issuecomment-514866924
 
 
   ## CI report:
   
   * 69b88611a023533b59dee72fe56332d77ee93884 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541108)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #7262: [FLINK-10478] Kafka Producer wrongly formats % for transaction ID

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #7262: [FLINK-10478] Kafka Producer wrongly 
formats % for transaction ID
URL: https://github.com/apache/flink/pull/7262#issuecomment-510752916
 
 
   ## CI report:
   
   * 15f05f4c5791d9d42610099324e59057d26bd3ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/118882595)
   * 2fc680ca11665e05fe0a66f3a92ba5e5d2d734cc : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120541120)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
flinkbot commented on issue #9222: [FLINK-13409][python] Supported java UDFs in 
python API.
URL: https://github.com/apache/flink/pull/9222#issuecomment-514865689
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13409) Supported java UDFs in python API

2019-07-24 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13409?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-13409:
---
Labels: pull-request-available  (was: )

> Supported java UDFs in python API
> -
>
> Key: FLINK-13409
> URL: https://issues.apache.org/jira/browse/FLINK-13409
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Python
>Affects Versions: 1.9.0
>Reporter: sunjincheng
>Assignee: Dian Fu
>Priority: Critical
>  Labels: pull-request-available
>
> It's better to support java UDF in python API.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13410) Csv input format does not support LocalDate

2019-07-24 Thread Caizhi Weng (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng updated FLINK-13410:

Issue Type: Bug  (was: Test)

> Csv input format does not support LocalDate
> ---
>
> Key: FLINK-13410
> URL: https://issues.apache.org/jira/browse/FLINK-13410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>
> Csv input format is lacking parsers for LocalDate, LocalTime, etc. As 
> DataTypes.DATE now defaults to LocalDate, we should add these parsers for 
> user experience.
> A temporal workaround for the users is that, users can call 
> DataTypes.Timestamp().bridgeTo(java.sql.Timestamp) to use the old 
> SqlTimestamp converter.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] WeiZhong94 opened a new pull request #9222: [FLINK-13409][python] Supported java UDFs in python API.

2019-07-24 Thread GitBox
WeiZhong94 opened a new pull request #9222: [FLINK-13409][python] Supported 
java UDFs in python API.
URL: https://github.com/apache/flink/pull/9222
 
 
   ## What is the purpose of the change
   
   *This pull request aimed to support registering Java UDFs and using Java 
UDFs in Python Table API.*
   
   
   ## Brief change log
   
 - *Add `register_java_function()` and `list_user_defined_functions()` in 
`TableEnvironment` class to support registering Java UDFs.*
 - *Add `join_lateral()` and `left_outer_join_lateral()` to support using 
UDTFs in python Table API.*
 - *Fix incorrect documentation of `joinLateral` interfaces. *
   
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as 
*test_correlate.py* and *test_table_environment_api.py*.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (python docs)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] leesf commented on a change in pull request #7262: [FLINK-10478] Kafka Producer wrongly formats % for transaction ID

2019-07-24 Thread GitBox
leesf commented on a change in pull request #7262: [FLINK-10478] Kafka Producer 
wrongly formats % for transaction ID
URL: https://github.com/apache/flink/pull/7262#discussion_r307088838
 
 

 ##
 File path: 
flink-connectors/flink-connector-kafka-0.11/src/main/java/org/apache/flink/streaming/connectors/kafka/internal/TransactionalIdsGenerator.java
 ##
 @@ -55,7 +55,7 @@ public TransactionalIdsGenerator(
checkArgument(safeScaleDownFactor > 0);
checkArgument(subtaskIndex >= 0);
 
-   this.prefix = checkNotNull(prefix);
+   this.prefix = checkNotNull(prefix).replaceAll("%", "%%");
 
 Review comment:
   @becketqin Updated this PR to address your comments. Please take another 
look.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-13410) Csv input format does not support LocalDate

2019-07-24 Thread Caizhi Weng (JIRA)
Caizhi Weng created FLINK-13410:
---

 Summary: Csv input format does not support LocalDate
 Key: FLINK-13410
 URL: https://issues.apache.org/jira/browse/FLINK-13410
 Project: Flink
  Issue Type: Test
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Reporter: Caizhi Weng


Csv input format is lacking parsers for LocalDate, LocalTime, etc. As 
DataTypes.DATE now defaults to LocalDate, we should add these parsers for user 
experience.

A temporal workaround for the users is that, users can call 
DataTypes.Timestamp().bridgeTo(java.sql.Timestamp) to use the old SqlTimestamp 
converter.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9141: [FLINK-12249][table] Fix type equivalence check problems for Window Aggregates

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9141: [FLINK-12249][table] Fix type 
equivalence check problems for Window Aggregates
URL: https://github.com/apache/flink/pull/9141#issuecomment-512181311
 
 
   ## CI report:
   
   * 1af43966364d445984524fd3bf51a0f3b8e75cbe : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119443400)
   * d7823627d50063f634d1ad6e43f470947a54daed : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119585913)
   * caf138832a23ecc7a3f72ffef760e5a3b070e0f1 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766257)
   * c6cd8ac0437314bf69074625d1a81dc72c664540 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120390913)
   * 5e29a7430750ece078ecf435938dde4e1a21366c : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120540112)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] hequn8128 commented on issue #9141: [FLINK-12249][table] Fix type equivalence check problems for Window Aggregates

2019-07-24 Thread GitBox
hequn8128 commented on issue #9141: [FLINK-12249][table] Fix type equivalence 
check problems for Window Aggregates
URL: https://github.com/apache/flink/pull/9141#issuecomment-514862245
 
 
   @godfreyhe Hi, thanks a lot for your review and suggestions. I have updated 
the PR according to your comments.
   
   Best, Hequn


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] lamber-ken commented on issue #9104: [HOTFIX][mvn] upgrade frontend-maven-plugin version to 1.7.5

2019-07-24 Thread GitBox
lamber-ken commented on issue #9104: [HOTFIX][mvn] upgrade 
frontend-maven-plugin version to 1.7.5
URL: https://github.com/apache/flink/pull/9104#issuecomment-514858367
 
 
   > I think this should be another problem when building frontend plugin, why 
it said cannot find `org.apa.http` in [line 
14](https://github.com/eirslett/frontend-maven-plugin/blob/e2162a03798b1980d1615af28ff3222fbb9d4012/frontend-plugin-core/src/main/java/com/github/eirslett/maven/plugins/frontend/lib/FileDownloader.java#L14)
 instead of `org.apache.http.HttpHost`?
   
   Hi, @Myasuka. This can already be explained that frontend-maven-plugin-1.6 
component will cause compilation failure.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Assigned] (FLINK-13409) Supported java UDFs in python API

2019-07-24 Thread sunjincheng (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13409?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

sunjincheng reassigned FLINK-13409:
---

Assignee: Dian Fu

> Supported java UDFs in python API
> -
>
> Key: FLINK-13409
> URL: https://issues.apache.org/jira/browse/FLINK-13409
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Python
>Affects Versions: 1.9.0
>Reporter: sunjincheng
>Assignee: Dian Fu
>Priority: Critical
>
> It's better to support java UDF in python API.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Created] (FLINK-13409) Supported java UDFs in python API

2019-07-24 Thread sunjincheng (JIRA)
sunjincheng created FLINK-13409:
---

 Summary: Supported java UDFs in python API
 Key: FLINK-13409
 URL: https://issues.apache.org/jira/browse/FLINK-13409
 Project: Flink
  Issue Type: Sub-task
  Components: API / Python
Affects Versions: 1.9.0
Reporter: sunjincheng


It's better to support java UDF in python API.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-6487) Remove JobManager local mode

2019-07-24 Thread TisonKun (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892285#comment-16892285
 ] 

TisonKun commented on FLINK-6487:
-

Thank [~StephanEwen]. I just found when execute the main method of Flink 
program directly we use MiniCluster.

> Remove JobManager local mode
> 
>
> Key: FLINK-6487
> URL: https://issues.apache.org/jira/browse/FLINK-6487
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Reporter: Stephan Ewen
>Priority: Major
>
> We should remove the "local" mode from the JobManager.
> Currently, the JobManager has the strange "local" mode where it also starts 
> an embedded Task Manager.
> I think that mode has caused confusion / problems:
>   - No TaskManagers can join the cluster
>   - TaskManagers do not support querable state
>   - It is redundant code to maintain
> At the same time, the mode does not help at all:
>   - The MiniCluster does not use that mode
>   - Starting from scripts, the {{start-cluster.sh}} works out of the box, 
> creating a proper local cluster, but in two processes, rather than one.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9210: [FLINK-12746][docs] Getting Started - DataStream Example Walkthrough

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9210: [FLINK-12746][docs] Getting Started - 
DataStream Example Walkthrough
URL: https://github.com/apache/flink/pull/9210#issuecomment-514437706
 
 
   ## CI report:
   
   * 5eb979da047c442c0205464c92b5bd9ee3a740dc : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120299964)
   * d7bf53a30514664925357bd5817305a02553d0a3 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120506936)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] Get SlotManager.failUnfulfillableRequest in main thread of StandaloneResourceManager for verification

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] 
Get SlotManager.failUnfulfillableRequest in main thread of 
StandaloneResourceManager for verification
URL: https://github.com/apache/flink/pull/9126#issuecomment-511719014
 
 
   ## CI report:
   
   * 910d6b067ec645e29edc9b05d324d9ca5a136a0b : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119428210)
   * e83800d72203e3b7221407b574c9ceacca473e84 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441098)
   * 56028391b7ede29b45ce04cb06d3d290185212bd : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120475864)
   * 8c0f1ea5c943a59b8d21bbd5b15c1176b45bfdf9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120479088)
   * ef8f74e06dbde2819c19d8f1dbb15bb6c58dc214 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120501596)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #8903: [FLINK-12747][docs] Getting Started - 
Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#issuecomment-510464651
 
 
   ## CI report:
   
   * b2821a6ae97fd943f3a66b672e85fbd2374126c4 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/118909729)
   * 0699f7e5f2240a4a1bc44c15f08e6a1df47d3b01 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119054579)
   * 509e634257496dd2d8d42d512901f5eb46a82c50 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119406891)
   * 7579df06b6a0bf799e8a9c2bcb09984bf52c8e8c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441302)
   * ccb9dc29d4755d0a6c4596e08743b38615eb276a : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120480063)
   * 1b976f30a689d9bdbf65513f034b2954bfb91468 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120494302)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #8903: [FLINK-12747][docs] Getting Started - 
Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#issuecomment-510464651
 
 
   ## CI report:
   
   * b2821a6ae97fd943f3a66b672e85fbd2374126c4 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/118909729)
   * 0699f7e5f2240a4a1bc44c15f08e6a1df47d3b01 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119054579)
   * 509e634257496dd2d8d42d512901f5eb46a82c50 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119406891)
   * 7579df06b6a0bf799e8a9c2bcb09984bf52c8e8c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441302)
   * ccb9dc29d4755d0a6c4596e08743b38615eb276a : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120480063)
   * 1b976f30a689d9bdbf65513f034b2954bfb91468 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120494302)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] Get SlotManager.failUnfulfillableRequest in main thread of StandaloneResourceManager for verification

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] 
Get SlotManager.failUnfulfillableRequest in main thread of 
StandaloneResourceManager for verification
URL: https://github.com/apache/flink/pull/9126#issuecomment-511719014
 
 
   ## CI report:
   
   * 910d6b067ec645e29edc9b05d324d9ca5a136a0b : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119428210)
   * e83800d72203e3b7221407b574c9ceacca473e84 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441098)
   * 56028391b7ede29b45ce04cb06d3d290185212bd : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120475864)
   * 8c0f1ea5c943a59b8d21bbd5b15c1176b45bfdf9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120479088)
   * ef8f74e06dbde2819c19d8f1dbb15bb6c58dc214 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120501596)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] Get SlotManager.failUnfulfillableRequest in main thread of StandaloneResourceManager for verification

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] 
Get SlotManager.failUnfulfillableRequest in main thread of 
StandaloneResourceManager for verification
URL: https://github.com/apache/flink/pull/9126#issuecomment-511719014
 
 
   ## CI report:
   
   * 910d6b067ec645e29edc9b05d324d9ca5a136a0b : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119428210)
   * e83800d72203e3b7221407b574c9ceacca473e84 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441098)
   * 56028391b7ede29b45ce04cb06d3d290185212bd : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120475864)
   * 8c0f1ea5c943a59b8d21bbd5b15c1176b45bfdf9 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120479088)
   * ef8f74e06dbde2819c19d8f1dbb15bb6c58dc214 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120501596)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9210: [FLINK-12746][docs] Getting Started - DataStream Example Walkthrough

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9210: [FLINK-12746][docs] Getting Started - 
DataStream Example Walkthrough
URL: https://github.com/apache/flink/pull/9210#issuecomment-514437706
 
 
   ## CI report:
   
   * 5eb979da047c442c0205464c92b5bd9ee3a740dc : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120299964)
   * d7bf53a30514664925357bd5817305a02553d0a3 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120506936)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] 
Getting Started - Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#discussion_r306980051
 
 

 ##
 File path: docs/getting-started/walkthroughs/table_api.md
 ##
 @@ -0,0 +1,445 @@
+---
+title: "Table API"
+nav-id: tableapiwalkthrough
+nav-title: 'Table API'
+nav-parent_id: walkthroughs
+nav-pos: 1
+---
+
+
+Apache Flink offers a Table API as a unified, relational API for batch and 
stream processing, i.e., queries are executed with the same semantics on 
unbounded, real-time streams or bounded, recorded streams and produce the same 
results.
+The Table API in Flink is commonly used to ease the definition of data 
analytics, data pipelining, and ETL applications.
+
+* This will be replaced by the TOC
+{:toc}
+
+## What Will You Be Building? 
+
+In this tutorial, you'll learn how to build a continuous ETL pipeline for 
tracking financial transactions by account over time.
+You will start by building your report as a nightly batch job, and then 
migrate to a streaming pipeline.
+
+## Prerequisites
+
+This walkthrough assumes that you have some familiarity with Java or Scala, 
but you should be able to follow along even if you're coming from a different 
programming language.
+It also assumes that you're familiar with basic relational concepts such as 
`SELECT` and `GROUP BY` clauses. 
+
+## Help, I’m Stuck! 
+
+If you get stuck, check out the [community support 
resources](https://flink.apache.org/community.html).
+In particular, Apache Flink's [user mailing 
list](https://flink.apache.org/community.html#mailing-lists) is consistently 
ranked as one of the most active of any Apache project and a great way to get 
help quickly. 
+
+## How To Follow Along
+
+If you want to follow along, you will require a computer with: 
+
+* Java 8 
+* Maven 
+
+There is also a provided Flink Maven Archetype to create a skeleton project 
with all the necessary dependencies quickly.
+
+
+
+{% highlight bash %}
+$ mvn archetype:generate \
+-DarchetypeGroupId=org.apache.flink \
+-DarchetypeArtifactId=flink-walkthrough-table-java \{% unless 
site.is_stable %}
+
-DarchetypeCatalog=https://repository.apache.org/content/repositories/snapshots/
 \{% endunless %}
+-DarchetypeVersion={{ site.version }} \
+-DgroupId=spend-report \
+-DartifactId=spend-report \
+-Dversion=0.1 \
+-Dpackage=spendreport \
+-DinteractiveMode=false
+{% endhighlight %}
+
+
+{% highlight bash %}
+$ mvn archetype:generate \
+-DarchetypeGroupId=org.apache.flink \
+-DarchetypeArtifactId=flink-walkthrough-table-scala \{% unless 
site.is_stable %}
+
-DarchetypeCatalog=https://repository.apache.org/content/repositories/snapshots/
 \{% endunless %}
+-DarchetypeVersion={{ site.version }} \
+-DgroupId=spend-report \
+-DartifactId=spend-report \
+-Dversion=0.1 \
+-Dpackage=spendreport \
+-DinteractiveMode=false
+{% endhighlight %}
+
+
+
+{% unless site.is_stable %}
+
+Note: For Maven 3.0 or higher, it is no longer possible to specify 
the repository (-DarchetypeCatalog) via the commandline. If you wish to use the 
snapshot repository, you need to add a repository entry to your settings.xml. 
For details about this change, please refer to http://maven.apache.org/archetype/maven-archetype-plugin/archetype-repository.html;>Maven
 official document
+
+{% endunless %}
+
+You can edit the `groupId`, `artifactId` and `package` if you like. With the 
above parameters,
+Maven will create a project with all the dependencies to complete this 
tutorial.
+After importing the project into your editor, you will see a file following 
code. 
+
+
+
+{% highlight java %}
+ExecutionEnvironment env   = ExecutionEnvironment.getExecutionEnvironment();
+BatchTableEnvironment tEnv = BatchTableEnvironment.create(env);
+
+tEnv.registerTableSource("transactions", new BoundedTransactionTableSource());
+tEnv.registerTableSink("spend_report", new SpendReportTableSink());
+tEnv.registerFunction("truncateDateToHour", new TruncateDateToHour());
+
+tEnv
+   .scan("transactions")
+   .insertInto("spend_report");
+
+env.execute("Spend Report");
+{% endhighlight %}
+
+
+
+{% highlight scala %}
+val env  = ExecutionEnvironment.getExecutionEnvironment
+val tEnv = BatchTableEnvironment.create(env)
+
+tEnv.registerTableSource("transactions", new BoundedTransactionTableSource)
+tEnv.registerTableSink("spend_report", new SpendReportTableSink)
+
+val truncateDateToHour = new TruncateDateToHour
+
+tEnv
+   .scan("transactions")
+   .insertInto("spend_report")
+
+env.execute("Spend Report")
+{% endhighlight %}
+
+
+
+Let's break down this code by component. 
+
+## Breaking Down The Code
+
+ The Execution Environment
+
+The first two lines set up your `ExecutionEnvironment`.
+The execution environment is how you can set properties for your Job, specify 
whether you are writing a batch or streaming 

[GitHub] [flink] sjwiesman commented on issue #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
sjwiesman commented on issue #8903: [FLINK-12747][docs] Getting Started - Table 
API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#issuecomment-514765009
 
 
   @NicoK I believe I have addressed all of your comments


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9203: [FLINK-13375][table-api] Improve config names in ExecutionConfigOptions and OptimizerConfigOptions

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9203: [FLINK-13375][table-api] Improve 
config names in ExecutionConfigOptions and OptimizerConfigOptions
URL: https://github.com/apache/flink/pull/9203#issuecomment-514046368
 
 
   ## CI report:
   
   * f5e680b52e6a85e85642fc22a41724c5a452505c : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120109078)
   * c1388ab2867ad134b2300ccad6ca519eff547ccb : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120216941)
   * b70392e5d7337da2a416316068397f92925a0d7e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120455507)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] Get SlotManager.failUnfulfillableRequest in main thread of StandaloneResourceManager for verification

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9126: [FLINK-13407][tests][coordination] 
Get SlotManager.failUnfulfillableRequest in main thread of 
StandaloneResourceManager for verification
URL: https://github.com/apache/flink/pull/9126#issuecomment-511719014
 
 
   ## CI report:
   
   * 910d6b067ec645e29edc9b05d324d9ca5a136a0b : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119428210)
   * e83800d72203e3b7221407b574c9ceacca473e84 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441098)
   * 56028391b7ede29b45ce04cb06d3d290185212bd : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120475864)
   * 8c0f1ea5c943a59b8d21bbd5b15c1176b45bfdf9 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120479088)
   * ef8f74e06dbde2819c19d8f1dbb15bb6c58dc214 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120501596)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9221: [FLINK-13376][datastream] ContinuousFileReaderOperator should respect…

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9221: [FLINK-13376][datastream] 
ContinuousFileReaderOperator should respect…
URL: https://github.com/apache/flink/pull/9221#issuecomment-514645431
 
 
   ## CI report:
   
   * 05fe281c79ab4bd2646ec949dcbbf0c2af9fec6e : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120456728)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9220: [FLINK-13403] [table] Correct package name after relocation (for release-1.9)

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9220: [FLINK-13403] [table] Correct package 
name after relocation (for release-1.9)
URL: https://github.com/apache/flink/pull/9220#issuecomment-514610294
 
 
   ## CI report:
   
   * 8d78497611d6907e2633d67f1791a8ae9db010ef : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120441818)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #8903: [FLINK-12747][docs] Getting Started - 
Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#issuecomment-510464651
 
 
   ## CI report:
   
   * b2821a6ae97fd943f3a66b672e85fbd2374126c4 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/118909729)
   * 0699f7e5f2240a4a1bc44c15f08e6a1df47d3b01 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119054579)
   * 509e634257496dd2d8d42d512901f5eb46a82c50 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119406891)
   * 7579df06b6a0bf799e8a9c2bcb09984bf52c8e8c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441302)
   * ccb9dc29d4755d0a6c4596e08743b38615eb276a : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120480063)
   * 1b976f30a689d9bdbf65513f034b2954bfb91468 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/120494302)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9022: [FLINK-13127] Fix --yarnship classpath

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9022: [FLINK-13127] Fix --yarnship classpath
URL: https://github.com/apache/flink/pull/9022#issuecomment-512166531
 
 
   ## CI report:
   
   * 277d92daa0e31cc2fced052a1b311cdba0f567e1 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119436815)
   * e1cd4a223022fd9a130dec570e87ec5d161d3173 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120448921)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-6487) Remove JobManager local mode

2019-07-24 Thread Stephan Ewen (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892039#comment-16892039
 ] 

Stephan Ewen commented on FLINK-6487:
-

The mini cluster is also used behind the Local[Stream]ExecutionEnvironment, so 
for examples, testing, debugging, etc.
It is definitely meant to be used by users, not just by tests.

> Remove JobManager local mode
> 
>
> Key: FLINK-6487
> URL: https://issues.apache.org/jira/browse/FLINK-6487
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Reporter: Stephan Ewen
>Priority: Major
>
> We should remove the "local" mode from the JobManager.
> Currently, the JobManager has the strange "local" mode where it also starts 
> an embedded Task Manager.
> I think that mode has caused confusion / problems:
>   - No TaskManagers can join the cluster
>   - TaskManagers do not support querable state
>   - It is redundant code to maintain
> At the same time, the mode does not help at all:
>   - The MiniCluster does not use that mode
>   - Starting from scripts, the {{start-cluster.sh}} works out of the box, 
> creating a proper local cluster, but in two processes, rather than one.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13306) flink-examples-streaming-gcp-pubsub is missing NOTICE

2019-07-24 Thread Stephan Ewen (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13306?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892038#comment-16892038
 ] 

Stephan Ewen commented on FLINK-13306:
--

Even more reason to remove it?

> flink-examples-streaming-gcp-pubsub is missing NOTICE
> -
>
> Key: FLINK-13306
> URL: https://issues.apache.org/jira/browse/FLINK-13306
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Google Cloud PubSub, Examples
>Affects Versions: 1.9.0
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Blocker
>
> The pubsub example is bundling various dependencies but is missing a NOTICE 
> file.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] azagrebin commented on a change in pull request #9175: [FLINK-12038] [test] fix YARNITCase random fail

2019-07-24 Thread GitBox
azagrebin commented on a change in pull request #9175: [FLINK-12038] [test] fix 
YARNITCase random fail
URL: https://github.com/apache/flink/pull/9175#discussion_r306932028
 
 

 ##
 File path: 
flink-yarn-tests/src/test/java/org/apache/flink/yarn/YarnTestBase.java
 ##
 @@ -573,6 +576,23 @@ public static int getRunningContainers() {
return count;
}
 
+   void waitUntilApplicationFinished(ApplicationId applicationId, Duration 
timeout) throws Exception {
+   Deadline deadline = Deadline.now().plus(timeout);
+   while (true) {
 
 Review comment:
   I would prefer
   ```
   YarnApplicationState state;
   do {
  state = 
yarnClient.getApplicationReport(applicationId).getYarnApplicationState();
  .
   } while (state != YarnApplicationState.FINISHED); 
   ```


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] azagrebin commented on a change in pull request #9175: [FLINK-12038] [test] fix YARNITCase random fail

2019-07-24 Thread GitBox
azagrebin commented on a change in pull request #9175: [FLINK-12038] [test] fix 
YARNITCase random fail
URL: https://github.com/apache/flink/pull/9175#discussion_r306933213
 
 

 ##
 File path: 
flink-yarn-tests/src/test/java/org/apache/flink/yarn/YarnTestBase.java
 ##
 @@ -573,6 +576,23 @@ public static int getRunningContainers() {
return count;
}
 
+   void waitUntilApplicationFinished(ApplicationId applicationId, Duration 
timeout) throws Exception {
+   Deadline deadline = Deadline.now().plus(timeout);
+   while (true) {
+   YarnApplicationState state = 
yarnClient.getApplicationReport(applicationId).getYarnApplicationState();
+   if (state == YarnApplicationState.FINISHED) {
+   break;
+   } else if (state == YarnApplicationState.FAILED || 
state == YarnApplicationState.KILLED) {
+   Assert.fail("Application became FAILED or 
KILLED while expecting FINISHED");
+   } else {
+   sleep(sleepIntervalInMS);
+   }
+   if (deadline.isOverdue()) {
+   Assert.fail("Application didn't finish before 
timeout");
 
 Review comment:
   I think we should move the app killing from the finally block of 
`YARNITCase. waitUntilApplicationFinished` to this timeout case, right before 
`Assert.fail("Application didn't finish before timeout");`. As we need killing 
only if the app hangs but we do not need if it finished successfully.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9212: [FLINK-13338][table-api] Sql conformance is hard to config in TableConfig

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9212: [FLINK-13338][table-api] Sql 
conformance is hard to config in TableConfig
URL: https://github.com/apache/flink/pull/9212#issuecomment-514486972
 
 
   ## CI report:
   
   * 68443427c89cca6ef5c1c39685c57c9884e101af : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120314731)
   * 51463ef8d90e7a3398b59c1efcf6ce713aaddb03 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120315489)
   * c839ffc015637729a586b0c78f61c54bad21b299 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120315878)
   * 39df37c7b430ed295b5b55a7b21e677aafd5413d : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/120426548)
   * 254e1cdc74f8d2e6d68f59606ee0b7867132e00d : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120448887)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13245) Network stack is leaking files

2019-07-24 Thread Stephan Ewen (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13245?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892023#comment-16892023
 ] 

Stephan Ewen commented on FLINK-13245:
--

An observation on the current state of the implementation of 
{{ReleaseOnConsumptionResultPartition}}

We have a weird inconsistent situation where we
  1. have a count down on the ResultPartition level about the 
{{notifySubpartitionConsumed()}} calls from the subpartitions, to release the 
partition when all subpartitions are released.
  2. have results that can have multiple readers/views and can hence receive 
multiple release/consumed calls.

This can probably lead to counting two notifications from one subpartition, and 
then releasing too early, unless there is super careful accounting when to 
notify about consumption and when not to.
This careful accounting seems super fragile to me, given the state of the netty 
stack. I expect that we will have issues were we either notify too often (early 
release) or not often enough (lingering files).

I would suggest to do the following: 
  - Change {{ReleaseOnConsumptionResultPartition}} to have a flag per 
subpartition that tracks whether there was a {{notifySubpartitionConsumed()}} 
call or not, so that multiple calls are idempotent.
  - Always send {{notifySubpartitionConsumed()}} calls with every 
{{releaseAllResources()}} call.

> Network stack is leaking files
> --
>
> Key: FLINK-13245
> URL: https://issues.apache.org/jira/browse/FLINK-13245
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.9.0
>Reporter: Chesnay Schepler
>Assignee: zhijiang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> There's file leak in the network stack / shuffle service.
> When running the {{SlotCountExceedingParallelismTest}} on Windows a large 
> number of {{.channel}} files continue to reside in a 
> {{flink-netty-shuffle-XXX}} directory.
> From what I've gathered so far these files are still being used by a 
> {{BoundedBlockingSubpartition}}. The cleanup logic in this class uses 
> ref-counting to ensure we don't release data while a reader is still present. 
> However, at the end of the job this count has not reached 0, and thus nothing 
> is being released.
> The same issue is also present on the {{ResultPartition}} level; the 
> {{ReleaseOnConsumptionResultPartition}} also are being released while the 
> ref-count is greater than 0.
> Overall it appears like there's some issue with the notifications for 
> partitions being consumed.
> It is feasible that this issue has recently caused issues on Travis where the 
> build were failing due to a lack of disk space.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] azagrebin edited a comment on issue #9126: [FLINK-13407][tests][coordination] Get SlotManager.failUnfulfillableRequest in main thread of StandaloneResourceManager for verification

2019-07-24 Thread GitBox
azagrebin edited a comment on issue #9126: [FLINK-13407][tests][coordination] 
Get SlotManager.failUnfulfillableRequest in main thread of 
StandaloneResourceManager for verification
URL: https://github.com/apache/flink/pull/9126#issuecomment-514722375
 
 
   Thanks for the review @tillrohrmann I have addressed the comments
   
   I think we can merge this PR separately. I factored out it into a separate 
[JIRA task](https://issues.apache.org/jira/browse/FLINK-13407) as a one step to 
fix the original [test instability 
FLINK-13242](https://issues.apache.org/jira/browse/FLINK-13242). Could you 
assign the task to me?
   
   We can address the latest finding in a separate [JIRA 
task](https://issues.apache.org/jira/browse/FLINK-13408). Could you assign the 
task to @xintongsong? 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] azagrebin commented on issue #9126: [FLINK-13407][tests][coordination] Get SlotManager.failUnfulfillableRequest in main thread of StandaloneResourceManager for verification

2019-07-24 Thread GitBox
azagrebin commented on issue #9126: [FLINK-13407][tests][coordination] Get 
SlotManager.failUnfulfillableRequest in main thread of 
StandaloneResourceManager for verification
URL: https://github.com/apache/flink/pull/9126#issuecomment-514722375
 
 
   Thanks for the review @tillrohrmann I have addressed the comments
   
   I think we can merge this PR separately. I factored out in a separate [JIRA 
task](https://issues.apache.org/jira/browse/FLINK-13407) as a one step to fix 
the original [test instability 
FLINK-13242](https://issues.apache.org/jira/browse/FLINK-13242). Could you 
assign the task to me?
   
   We can address the latest finding in a separate [JIRA 
task](https://issues.apache.org/jira/browse/FLINK-13408). Could you assign the 
task to @xintongsong? 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (FLINK-13245) Network stack is leaking files

2019-07-24 Thread Stephan Ewen (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13245?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892005#comment-16892005
 ] 

Stephan Ewen edited comment on FLINK-13245 at 7/24/19 5:16 PM:
---

Thanks for this discussion. I commented on the PR suggestion to always call 
{{notifySubpartitionConsumed()}} when releasing a reader.

My suggestion for Flink 1.10 would be:
  - Drop {{notifySubpartitionConsumed()}} completely
  - Drop the {{ReleaseOnConsumptionResultPartition}}
  - For bounded blocking partitions, the release happens always from the 
scheduler (no {{JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION}} any 
more)
  - Pipelined subpartitions are released when the one and only reader/view is 
released. There can be no further reader, so might as well immediate release 
it. The result partition as whole is released when all subpartitions are 
released.



was (Author: stephanewen):
Thanks for this discussion. I commented on the PR suggestion to always call 
{{notifySubpartitionConsumed()}} when releasing a reader.

My suggestion for Flink 1.10 would be:
  - Drop {{notifySubpartitionConsumed()}} completely
  - Drop the {{ReleaseOnConsumptionResultPartition}}
  - For bounded blocking partitions, the release happens always from the 
scheduler (no {{JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION}} any 
more)
  - Pipelined partitions are released when the one and only reader/view is 
released. There can be no further reader, so might as well immediate release it.

> Network stack is leaking files
> --
>
> Key: FLINK-13245
> URL: https://issues.apache.org/jira/browse/FLINK-13245
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.9.0
>Reporter: Chesnay Schepler
>Assignee: zhijiang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> There's file leak in the network stack / shuffle service.
> When running the {{SlotCountExceedingParallelismTest}} on Windows a large 
> number of {{.channel}} files continue to reside in a 
> {{flink-netty-shuffle-XXX}} directory.
> From what I've gathered so far these files are still being used by a 
> {{BoundedBlockingSubpartition}}. The cleanup logic in this class uses 
> ref-counting to ensure we don't release data while a reader is still present. 
> However, at the end of the job this count has not reached 0, and thus nothing 
> is being released.
> The same issue is also present on the {{ResultPartition}} level; the 
> {{ReleaseOnConsumptionResultPartition}} also are being released while the 
> ref-count is greater than 0.
> Overall it appears like there's some issue with the notifications for 
> partitions being consumed.
> It is feasible that this issue has recently caused issues on Travis where the 
> build were failing due to a lack of disk space.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Created] (FLINK-13408) Schedule StandaloneResourceManager.setFailUnfulfillableRequest whenever the leadership is acquired

2019-07-24 Thread Andrey Zagrebin (JIRA)
Andrey Zagrebin created FLINK-13408:
---

 Summary: Schedule 
StandaloneResourceManager.setFailUnfulfillableRequest whenever the leadership 
is acquired
 Key: FLINK-13408
 URL: https://issues.apache.org/jira/browse/FLINK-13408
 Project: Flink
  Issue Type: Sub-task
  Components: Runtime / Coordination
Affects Versions: 1.9.0
Reporter: Andrey Zagrebin
 Fix For: 1.9.0, 1.10.0


We introduced _StandaloneResourceManager.__setFailUnfulfillableRequest_ to give 
some time to task executors to register the available slots before the slot 
requests can be checked whether they can be fulfilled or not. 
_setFailUnfulfillableRequest_ is scheduled now only once when the RM is 
initialised but the task executors will register themselves every time this RM 
gets the leadership. Hence, _setFailUnfulfillableRequest_ should be scheduled 
after each leader election.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] Myasuka commented on issue #9104: [HOTFIX][mvn] upgrade frontend-maven-plugin version to 1.7.5

2019-07-24 Thread GitBox
Myasuka commented on issue #9104: [HOTFIX][mvn] upgrade frontend-maven-plugin 
version to 1.7.5
URL: https://github.com/apache/flink/pull/9104#issuecomment-514717512
 
 
   I think this should be another problem when building frontend plugin, why it 
said cannot find `org.apa.http` in [line 
14](https://github.com/eirslett/frontend-maven-plugin/blob/e2162a03798b1980d1615af28ff3222fbb9d4012/frontend-plugin-core/src/main/java/com/github/eirslett/maven/plugins/frontend/lib/FileDownloader.java#L14)
 instead of `org.apache.http.HttpHost`?
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13245) Network stack is leaking files

2019-07-24 Thread Stephan Ewen (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13245?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16892005#comment-16892005
 ] 

Stephan Ewen commented on FLINK-13245:
--

Thanks for this discussion. I commented on the PR suggestion to always call 
{{notifySubpartitionConsumed()}} when releasing a reader.

My suggestion for Flink 1.10 would be:
  - Drop {{notifySubpartitionConsumed()}} completely
  - Drop the {{ReleaseOnConsumptionResultPartition}}
  - For bounded blocking partitions, the release happens always from the 
scheduler (no {{JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION}} any 
more)
  - Pipelined partitions are released when the one and only reader/view is 
released. There can be no further reader, so might as well immediate release it.

> Network stack is leaking files
> --
>
> Key: FLINK-13245
> URL: https://issues.apache.org/jira/browse/FLINK-13245
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.9.0
>Reporter: Chesnay Schepler
>Assignee: zhijiang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> There's file leak in the network stack / shuffle service.
> When running the {{SlotCountExceedingParallelismTest}} on Windows a large 
> number of {{.channel}} files continue to reside in a 
> {{flink-netty-shuffle-XXX}} directory.
> From what I've gathered so far these files are still being used by a 
> {{BoundedBlockingSubpartition}}. The cleanup logic in this class uses 
> ref-counting to ensure we don't release data while a reader is still present. 
> However, at the end of the job this count has not reached 0, and thus nothing 
> is being released.
> The same issue is also present on the {{ResultPartition}} level; the 
> {{ReleaseOnConsumptionResultPartition}} also are being released while the 
> ref-count is greater than 0.
> Overall it appears like there's some issue with the notifications for 
> partitions being consumed.
> It is feasible that this issue has recently caused issues on Travis where the 
> build were failing due to a lack of disk space.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] 
Getting Started - Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#discussion_r306917597
 
 

 ##
 File path: 
flink-walkthroughs/flink-walkthrough-table-scala/src/main/resources/archetype-resources/pom.xml
 ##
 @@ -0,0 +1,333 @@
+
+http://maven.apache.org/POM/4.0.0;
+xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd;>
+   4.0.0
+
+   ${groupId}
+   ${artifactId}
+   ${version}
+   jar
+
+   Flink Walkthrough Table
+   http://www.myorganization.org
+
+   
+   
+   apache.snapshots
+   Apache Development Snapshot Repository
+   
https://repository.apache.org/content/repositories/snapshots/
+   
+   false
+   
+   
+   true
+   
+   
+   
+
+   
+   
UTF-8
+   @project.version@
+   2.11
+   2.11.12
+   
+
+   
+   
+   org.apache.flink
+   
flink-walkthrough-common_${scala.binary.version}
+   ${flink.version}
+   
+
+   
+   
+   
+   org.apache.flink
+   
flink-scala_${scala.binary.version}
+   ${flink.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-streaming-scala_${scala.binary.version}
+   ${flink.version}
+   provided
+   
+
+   
+   
+   org.scala-lang
+   scala-library
+   ${scala.version}
+   provided
+   
+
+   
+   
+   org.apache.flink
+   
flink-table-api-java-bridge_${scala.binary.version}
+   ${flink.version}
+   provided
+   
 
 Review comment:
   Yes, actually `flink-walkthrough-common` should only depend on 
`flink-table-common` but there are some classes that haven't been ported yet. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] 
Getting Started - Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#discussion_r306917652
 
 

 ##
 File path: flink-walkthroughs/flink-walkthrough-common/pom.xml
 ##
 @@ -0,0 +1,71 @@
+
+
+http://maven.apache.org/POM/4.0.0;
+xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/maven-v4_0_0.xsd;>
+
+   4.0.0
+
+   
+   org.apache.flink
+   flink-walkthroughs
+   1.10-SNAPSHOT
+   ..
+   
+
+   
flink-walkthrough-common_${scala.binary.version}
+   flink-walkthrough-common
+
+   jar
+
+   
+   
+   org.apache.flink
+   flink-java
+   ${project.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-streaming-java_${scala.binary.version}
+   ${project.version}
+   provided
+   
+   
+   org.apache.flink
+   flink-table-common
+   ${project.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-table-api-java-bridge_${scala.binary.version}
+   ${project.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-table-planner_${scala.binary.version}
+   ${project.version}
+   provided
+   
 
 Review comment:
   Same as above


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] Getting Started - Table API Example Walkthrough

2019-07-24 Thread GitBox
sjwiesman commented on a change in pull request #8903: [FLINK-12747][docs] 
Getting Started - Table API Example Walkthrough
URL: https://github.com/apache/flink/pull/8903#discussion_r306916406
 
 

 ##
 File path: 
flink-walkthroughs/flink-walkthrough-table-scala/src/main/resources/archetype-resources/pom.xml
 ##
 @@ -0,0 +1,333 @@
+
+http://maven.apache.org/POM/4.0.0;
+xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd;>
+   4.0.0
+
+   ${groupId}
+   ${artifactId}
+   ${version}
+   jar
+
+   Flink Walkthrough Table
+   http://www.myorganization.org
+
+   
+   
+   apache.snapshots
+   Apache Development Snapshot Repository
+   
https://repository.apache.org/content/repositories/snapshots/
+   
+   false
+   
+   
+   true
+   
+   
+   
+
+   
+   
UTF-8
+   @project.version@
+   2.11
+   2.11.12
+   
+
+   
+   
+   org.apache.flink
+   
flink-walkthrough-common_${scala.binary.version}
+   ${flink.version}
+   
+
+   
+   
+   
+   org.apache.flink
+   
flink-scala_${scala.binary.version}
+   ${flink.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-streaming-scala_${scala.binary.version}
+   ${flink.version}
+   provided
+   
+
+   
+   
+   org.scala-lang
+   scala-library
+   ${scala.version}
+   provided
+   
+
+   
+   
+   org.apache.flink
+   
flink-table-api-java-bridge_${scala.binary.version}
+   ${flink.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-table-api-scala-bridge_${scala.binary.version}
+   ${flink.version}
+   provided
+   
+   
+   org.apache.flink
+   
flink-table-planner_${scala.binary.version}
+   ${flink.version}
+   provided
+   
 
 Review comment:
   The other option is to copy `flink-table-uber` from `/opt` to '/lib' which I 
think is a cleaner solution. I will add a note to the walkthrough. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] StephanEwen edited a comment on issue #9132: [FLINK-13245][network] Fix the bug of file resource leak while canceling partition request

2019-07-24 Thread GitBox
StephanEwen edited a comment on issue #9132: [FLINK-13245][network] Fix the bug 
of file resource leak while canceling partition request
URL: https://github.com/apache/flink/pull/9132#issuecomment-514705841
 
 
   I am wondering why we distinguish cases where we release with/without 
consumption notification.
   To me, all release should include a consumption notification - similar to 
what @azagrebin raised in the JIRA discussion.
   
   The way I understand it, `notifySubpartitionConsumed()` has a semantic of "a 
consumption attempt finished", irrespective of whether it was complete, 
incomplete, cancelled, etc. All release cases indicate that end of a 
consumption attempt.
   
   I understand that this is not exactly the behavior we had before, but if the 
above is true, the behavior before was arguably incorrect and we should fix 
this now.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9214: [FLINK-13397] Remove RollingSink

2019-07-24 Thread GitBox
flinkbot edited a comment on issue #9214: [FLINK-13397] Remove RollingSink
URL: https://github.com/apache/flink/pull/9214#issuecomment-514554707
 
 
   ## CI report:
   
   * 408e5ac4cb1a6b69b400a3a80147c3202ecc62ab : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120408031)
   * 76b9f95823bb8b5cb01cc13a4d02913e7612cce4 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/120439971)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


  1   2   3   4   5   >