[jira] [Created] (FLINK-19083) Remove deprecated DataStream#split

2020-08-27 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-19083:


 Summary: Remove deprecated DataStream#split
 Key: FLINK-19083
 URL: https://issues.apache.org/jira/browse/FLINK-19083
 Project: Flink
  Issue Type: Sub-task
  Components: API / DataStream
Reporter: Dawid Wysakowicz






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19035) Remove deprecated DataStream#fold() method and all related classes

2020-08-27 Thread Dawid Wysakowicz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19035?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Wysakowicz updated FLINK-19035:
-
Fix Version/s: 2.0.0

> Remove deprecated DataStream#fold() method and all related classes
> --
>
> Key: FLINK-19035
> URL: https://issues.apache.org/jira/browse/FLINK-19035
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Reporter: Dawid Wysakowicz
>Priority: Major
> Fix For: 2.0.0
>
>
> We should also agree if we want to remove those in 1.x version already or do 
> we want to wait for 2.0.
> We should remove DataStream#fold and all related classes and methods such as 
> FoldFunction, FoldingState, FoldingStateDescriptor ... 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19083) Remove deprecated DataStream#split

2020-08-27 Thread Dawid Wysakowicz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Wysakowicz updated FLINK-19083:
-
Fix Version/s: 2.0.0

> Remove deprecated DataStream#split
> --
>
> Key: FLINK-19083
> URL: https://issues.apache.org/jira/browse/FLINK-19083
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Reporter: Dawid Wysakowicz
>Priority: Major
> Fix For: 2.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13263:
URL: https://github.com/apache/flink/pull/13263#issuecomment-681892463


   
   ## CI report:
   
   * 4f0f42441fadc0961dbdecc9f1cf5ced660387d4 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5946)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13273:
URL: https://github.com/apache/flink/pull/13273#issuecomment-682353427


   
   ## CI report:
   
   * 0134fa06171742eb8ae840b2aced9530232bdb04 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5955)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13274: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…

2020-08-27 Thread GitBox


flinkbot commented on pull request #13274:
URL: https://github.com/apache/flink/pull/13274#issuecomment-682362109


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 7835ad0187076a7dae2ee7e8806337f1b3db01c1 (Fri Aug 28 
06:55:05 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] lirui-apache opened a new pull request #13274: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…

2020-08-27 Thread GitBox


lirui-apache opened a new pull request #13274:
URL: https://github.com/apache/flink/pull/13274


   …f partition value contains special characters
   
   
   
   ## What is the purpose of the change
   
   Fix the issue that HiveCatalog fails to get partition column stats if 
partition value contains special characters
   
   
   ## Brief change log
   
 - Escape partition name when get/alter partition column stats
 - Update test to cover the use case
   
   
   ## Verifying this change
   
   Updated test case
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? NA
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] XComp commented on a change in pull request #13251: [FLINK-14435] Added memory configuration to TaskManagers REST endpoint

2020-08-27 Thread GitBox


XComp commented on a change in pull request #13251:
URL: https://github.com/apache/flink/pull/13251#discussion_r478866388



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/taskexecutor/TaskExecutorMemoryConfiguration.java
##
@@ -0,0 +1,333 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.taskexecutor;
+
+import org.apache.flink.configuration.ConfigOption;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.MemorySize;
+
+import org.apache.flink.shaded.guava18.com.google.common.base.MoreObjects;
+import 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonCreator;
+import 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonInclude;
+import 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonProperty;
+
+import java.io.Serializable;
+import java.util.Objects;
+
+import static 
org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_HEAP_MEMORY;
+import static 
org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_OFF_HEAP_MEMORY;
+import static org.apache.flink.configuration.TaskManagerOptions.JVM_METASPACE;
+import static 
org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_FRACTION;
+import static 
org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MAX;
+import static 
org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MIN;
+import static 
org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_FRACTION;
+import static 
org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_SIZE;
+import static 
org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_FRACTION;
+import static 
org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MAX;
+import static 
org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MIN;
+import static 
org.apache.flink.configuration.TaskManagerOptions.TASK_HEAP_MEMORY;
+import static 
org.apache.flink.configuration.TaskManagerOptions.TASK_OFF_HEAP_MEMORY;
+import static 
org.apache.flink.configuration.TaskManagerOptions.TOTAL_PROCESS_MEMORY;
+
+/**
+ * TaskExecutorConfiguration collects the configuration of a TaskExecutor 
instance.
+ */
+public class TaskExecutorMemoryConfiguration implements Serializable {
+
+   public static final String FIELD_NAME_FRAMEWORK_HEAP = "frameworkHeap";
+   public static final String FIELD_NAME_TASK_HEAP = "taskHeap";
+
+   public static final String FIELD_NAME_FRAMEWORK_OFFHEAP = 
"frameworkOffHeap";
+   public static final String FIELD_NAME_TASK_OFFHEAP = "taskOffHeap";
+
+   public static final String FIELD_NAME_NETWORK_MIN = "networkMin";
+   public static final String FIELD_NAME_NETWORK_MAX = "networkMax";
+   public static final String FIELD_NAME_NETWORK_FRACTION = 
"networkFraction";
+
+   public static final String FIELD_NAME_MANAGED_TOTAL = "managedTotal";
+   public static final String FIELD_NAME_MANAGED_FRACTION = 
"managedFraction";
+
+   public static final String FIELD_NAME_METASPACE_MAX = "metaspaceMax";
+
+   public static final String FIELD_NAME_OVERHEAD_MIN = "overheadMin";
+   public static final String FIELD_NAME_OVERHEAD_MAX = "overheadMax";
+   public static final String FIELD_NAME_OVERHEAD_FRACTION = 
"overheadFraction";
+
+   public static final String FIELD_NAME_MEMORY_TOTAL = "memoryTotal";
+
+   @JsonProperty(FIELD_NAME_FRAMEWORK_HEAP)
+   @JsonInclude
+   private final Long frameworkHeap;
+
+   @JsonProperty(FIELD_NAME_TASK_HEAP)
+   private final Long taskHeap;
+
+   @JsonProperty(FIELD_NAME_FRAMEWORK_OFFHEAP)
+   private final Long frameworkOffHeap;
+
+   @JsonProperty(FIELD_NAME_TASK_OFFHEAP)
+   private final Long taskOffHeap;
+
+   @JsonProperty(FIELD_NAME_NETWORK_MIN)
+   private final Long networkMemoryMin;
+   @JsonProperty(FIELD_NAME_NETWORK_MAX)
+   private final Long networkMemoryMax;
+   @JsonProperty(FIELD_NAME_NETWORK_FRACTION)
+   private final Float networkMemoryFraction;
+
+   @JsonProperty(FIELD_NAME_MANAGED_TOTAL)
+   private final Long managedMemoryTotal;
+ 

[GitHub] [flink] XComp commented on a change in pull request #13251: [FLINK-14435] Added memory configuration to TaskManagers REST endpoint

2020-08-27 Thread GitBox


XComp commented on a change in pull request #13251:
URL: https://github.com/apache/flink/pull/13251#discussion_r478866224



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/taskexecutor/TaskExecutorMemoryConfiguration.java
##
@@ -0,0 +1,333 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.taskexecutor;
+
+import org.apache.flink.configuration.ConfigOption;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.MemorySize;
+
+import org.apache.flink.shaded.guava18.com.google.common.base.MoreObjects;
+import 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonCreator;
+import 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonInclude;
+import 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonProperty;
+
+import java.io.Serializable;
+import java.util.Objects;
+
+import static 
org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_HEAP_MEMORY;
+import static 
org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_OFF_HEAP_MEMORY;
+import static org.apache.flink.configuration.TaskManagerOptions.JVM_METASPACE;
+import static 
org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_FRACTION;
+import static 
org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MAX;
+import static 
org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MIN;
+import static 
org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_FRACTION;
+import static 
org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_SIZE;
+import static 
org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_FRACTION;
+import static 
org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MAX;
+import static 
org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MIN;
+import static 
org.apache.flink.configuration.TaskManagerOptions.TASK_HEAP_MEMORY;
+import static 
org.apache.flink.configuration.TaskManagerOptions.TASK_OFF_HEAP_MEMORY;
+import static 
org.apache.flink.configuration.TaskManagerOptions.TOTAL_PROCESS_MEMORY;
+
+/**
+ * TaskExecutorConfiguration collects the configuration of a TaskExecutor 
instance.
+ */
+public class TaskExecutorMemoryConfiguration implements Serializable {
+
+   public static final String FIELD_NAME_FRAMEWORK_HEAP = "frameworkHeap";
+   public static final String FIELD_NAME_TASK_HEAP = "taskHeap";
+
+   public static final String FIELD_NAME_FRAMEWORK_OFFHEAP = 
"frameworkOffHeap";
+   public static final String FIELD_NAME_TASK_OFFHEAP = "taskOffHeap";
+
+   public static final String FIELD_NAME_NETWORK_MIN = "networkMin";
+   public static final String FIELD_NAME_NETWORK_MAX = "networkMax";
+   public static final String FIELD_NAME_NETWORK_FRACTION = 
"networkFraction";
+
+   public static final String FIELD_NAME_MANAGED_TOTAL = "managedTotal";
+   public static final String FIELD_NAME_MANAGED_FRACTION = 
"managedFraction";
+
+   public static final String FIELD_NAME_METASPACE_MAX = "metaspaceMax";
+
+   public static final String FIELD_NAME_OVERHEAD_MIN = "overheadMin";
+   public static final String FIELD_NAME_OVERHEAD_MAX = "overheadMax";
+   public static final String FIELD_NAME_OVERHEAD_FRACTION = 
"overheadFraction";
+
+   public static final String FIELD_NAME_MEMORY_TOTAL = "memoryTotal";
+
+   @JsonProperty(FIELD_NAME_FRAMEWORK_HEAP)
+   @JsonInclude
+   private final Long frameworkHeap;
+
+   @JsonProperty(FIELD_NAME_TASK_HEAP)
+   private final Long taskHeap;
+
+   @JsonProperty(FIELD_NAME_FRAMEWORK_OFFHEAP)
+   private final Long frameworkOffHeap;
+
+   @JsonProperty(FIELD_NAME_TASK_OFFHEAP)
+   private final Long taskOffHeap;
+
+   @JsonProperty(FIELD_NAME_NETWORK_MIN)
+   private final Long networkMemoryMin;
+   @JsonProperty(FIELD_NAME_NETWORK_MAX)
+   private final Long networkMemoryMax;
+   @JsonProperty(FIELD_NAME_NETWORK_FRACTION)
+   private final Float networkMemoryFraction;
+
+   @JsonProperty(FIELD_NAME_MANAGED_TOTAL)
+   private final Long managedMemoryTotal;
+ 

[GitHub] [flink] tillrohrmann commented on a change in pull request #13265: [FLINK-19055] Wait less time for all memory GC in tests (MemoryManager#verifyEmpty)

2020-08-27 Thread GitBox


tillrohrmann commented on a change in pull request #13265:
URL: https://github.com/apache/flink/pull/13265#discussion_r478863140



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/memory/UnsafeMemoryBudget.java
##
@@ -36,16 +36,20 @@
  */
 class UnsafeMemoryBudget {
private static final int MAX_SLEEPS = 11; // 2^11 - 1 = (2 x 1024) - 1 
ms ~ 2 s total sleep duration
-   private static final int MAX_SLEEPS_VERIFY_EMPTY = 17; // 2^17 - 1 = 
(128 x 1024) - 1 ms ~ 2 min total sleep duration
+   static final int MAX_SLEEPS_VERIFY_EMPTY_FOR_TESTS = 10; // 2^10 - 1 = 
(1 x 1024) - 1 ms ~ 1 s total sleep duration

Review comment:
   Testing code should not be contained in production code.

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/memory/MemoryManager.java
##
@@ -89,12 +89,14 @@
 *
 * @param memorySize The total size of the off-heap memory to be 
managed by this memory manager.
 * @param pageSize The size of the pages handed out by the memory 
manager.
+* @param verifyEmptyWaitGcMaxSleeps defines how long to wait for GC of 
all allocated memory to check for memory leaks,
+*   see also {@link 
UnsafeMemoryBudget}.

Review comment:
   Hmm, the description says that it is the time to wait for GC but isn't 
is the number of exponential retries?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Issue Comment Deleted] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Nicholas Jiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nicholas Jiang updated FLINK-19064:
---
Comment: was deleted

(was: [~ZhuShang] Please pay attention to the defintion of method 
openInputFormat, which javadoc mentions that Resources should be allocated in 
this method. (e.g. database connections, cache, etc.).)

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.

2020-08-27 Thread GitBox


flinkbot commented on pull request #13273:
URL: https://github.com/apache/flink/pull/13273#issuecomment-682353427


   
   ## CI report:
   
   * 0134fa06171742eb8ae840b2aced9530232bdb04 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13272:
URL: https://github.com/apache/flink/pull/13272#issuecomment-682346986


   
   ## CI report:
   
   * 3ef9f728f6a8005a7f91846bf32df96ee718c626 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5952)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13268:
URL: https://github.com/apache/flink/pull/13268#issuecomment-682283788


   
   ## CI report:
   
   * 5357812df45c42860e95630891367a410f6bc34b Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5944)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19061) HiveCatalog fails to get partition column stats if partition value contains special characters

2020-08-27 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186297#comment-17186297
 ] 

Jingsong Lee commented on FLINK-19061:
--

master: f924a665b6f0bff815f87a221495b101b035315e

> HiveCatalog fails to get partition column stats if partition value contains 
> special characters
> --
>
> Key: FLINK-19061
> URL: https://issues.apache.org/jira/browse/FLINK-19061
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi merged pull request #13264: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…

2020-08-27 Thread GitBox


JingsongLi merged pull request #13264:
URL: https://github.com/apache/flink/pull/13264


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Assigned] (FLINK-19061) HiveCatalog fails to get partition column stats if partition value contains special characters

2020-08-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee reassigned FLINK-19061:


Assignee: Rui Li

> HiveCatalog fails to get partition column stats if partition value contains 
> special characters
> --
>
> Key: FLINK-19061
> URL: https://issues.apache.org/jira/browse/FLINK-19061
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi commented on pull request #13264: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…

2020-08-27 Thread GitBox


JingsongLi commented on pull request #13264:
URL: https://github.com/apache/flink/pull/13264#issuecomment-682350350


   Can you create a PR for release-1.11?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-19042) HiveTableSourceITCase fails if object reuse is enabled

2020-08-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-19042.

  Assignee: Rui Li
Resolution: Fixed

master: 9e0964b81e502b266832fe2c271a1f8afdda5e5a

> HiveTableSourceITCase fails if object reuse is enabled
> --
>
> Key: FLINK-19042
> URL: https://issues.apache.org/jira/browse/FLINK-19042
> Project: Flink
>  Issue Type: Test
>  Components: Connectors / Hive, Tests
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Major
>  Labels: pull-request-available
>
> {{testNonPartitionStreamingSourceWithVectorizedReader}} fails because print 
> table sink cannot process {{ColumnarRowData}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19042) HiveTableSourceITCase fails if object reuse is enabled

2020-08-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-19042:
-
Fix Version/s: 1.12.0

> HiveTableSourceITCase fails if object reuse is enabled
> --
>
> Key: FLINK-19042
> URL: https://issues.apache.org/jira/browse/FLINK-19042
> Project: Flink
>  Issue Type: Test
>  Components: Connectors / Hive, Tests
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> {{testNonPartitionStreamingSourceWithVectorizedReader}} fails because print 
> table sink cannot process {{ColumnarRowData}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi merged pull request #13238: [FLINK-19042][hive] Remove print table sink from HiveTableSourceITCas…

2020-08-27 Thread GitBox


JingsongLi merged pull request #13238:
URL: https://github.com/apache/flink/pull/13238


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.

2020-08-27 Thread GitBox


flinkbot commented on pull request #13273:
URL: https://github.com/apache/flink/pull/13273#issuecomment-682347761


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 0134fa06171742eb8ae840b2aced9530232bdb04 (Fri Aug 28 
06:16:50 UTC 2020)
   
✅no warnings
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18801) Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section

2020-08-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18801?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18801:
---
Labels: pull-request-available  (was: )

> Add a "10 minutes to Table API" document under  the "Python API" -> "User 
> Guide" -> "Table API" section
> ---
>
> Key: FLINK-18801
> URL: https://issues.apache.org/jira/browse/FLINK-18801
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Wei Zhong
>Assignee: Wei Zhong
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19071) Some Hive window functions are not supported

2020-08-27 Thread Rui Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated FLINK-19071:
---
Description: 
Including the following:
{noformat}
FIRST_VALUE and LAST_VALUE that take two parameters
NTILE
CUME_DIST
PERCENT_RANK
{noformat}

  was:
Including the following:
{noformat}
FIRST_VALUE and LAST_VALUE that take two parameters
NTILE
CUME_DIST
{noformat}


> Some Hive window functions are not supported
> 
>
> Key: FLINK-19071
> URL: https://issues.apache.org/jira/browse/FLINK-19071
> Project: Flink
>  Issue Type: Task
>  Components: Connectors / Hive
>Reporter: Rui Li
>Priority: Major
>
> Including the following:
> {noformat}
> FIRST_VALUE and LAST_VALUE that take two parameters
> NTILE
> CUME_DIST
> PERCENT_RANK
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers

2020-08-27 Thread GitBox


flinkbot commented on pull request #13272:
URL: https://github.com/apache/flink/pull/13272#issuecomment-682346986


   
   ## CI report:
   
   * 3ef9f728f6a8005a7f91846bf32df96ee718c626 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] WeiZhong94 opened a new pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.

2020-08-27 Thread GitBox


WeiZhong94 opened a new pull request #13273:
URL: https://github.com/apache/flink/pull/13273


   ## What is the purpose of the change
   
   *This pull request adds a "10 minutes to Table API" document under the 
"Python API" -> "User Guide" -> "Table API" section.*
   
   
   ## Brief change log
   
 - *Add a "10 minutes to Table API" document under the "Python API" -> 
"User Guide" -> "Table API" section.*
   
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: ( no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (docs)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13271:
URL: https://github.com/apache/flink/pull/13271#issuecomment-682323547


   
   ## CI report:
   
   * dcfea8ba4f522e9cd1a87022eadb10438902f22c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5950)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13249: [FLINK-14087][datastream] Clone the StreamPartitioner to avoid being shared at runtime.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13249:
URL: https://github.com/apache/flink/pull/13249#issuecomment-680775587


   
   ## CI report:
   
   * 07e11ae0c700a4172a854ae4fd108b483bde6003 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5947)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19071) Some Hive window functions are not supported

2020-08-27 Thread Rui Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Li updated FLINK-19071:
---
Description: 
Including the following:
{noformat}
FIRST_VALUE and LAST_VALUE that take two parameters
NTILE
CUME_DIST
{noformat}

> Some Hive window functions are not supported
> 
>
> Key: FLINK-19071
> URL: https://issues.apache.org/jira/browse/FLINK-19071
> Project: Flink
>  Issue Type: Task
>  Components: Connectors / Hive
>Reporter: Rui Li
>Priority: Major
>
> Including the following:
> {noformat}
> FIRST_VALUE and LAST_VALUE that take two parameters
> NTILE
> CUME_DIST
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers

2020-08-27 Thread GitBox


flinkbot commented on pull request #13272:
URL: https://github.com/apache/flink/pull/13272#issuecomment-682341723


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 3ef9f728f6a8005a7f91846bf32df96ee718c626 (Fri Aug 28 
05:58:42 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] klion26 commented on a change in pull request #13225: [FLINK-18974][docs-zh]Translate the 'User-Defined Functions' page of "Application Development's DataStream API" into Chinese

2020-08-27 Thread GitBox


klion26 commented on a change in pull request #13225:
URL: https://github.com/apache/flink/pull/13225#discussion_r478844519



##
File path: docs/dev/user_defined_functions.zh.md
##
@@ -147,95 +153,77 @@ data.map (new RichMapFunction[String, Int] {
 
 
 
-Rich functions provide, in addition to the user-defined function (map,
-reduce, etc), four methods: `open`, `close`, `getRuntimeContext`, and
-`setRuntimeContext`. These are useful for parameterizing the function
-(see [Passing Parameters to Functions]({{ site.baseurl 
}}/dev/batch/index.html#passing-parameters-to-functions)),
-creating and finalizing local state, accessing broadcast variables (see
-[Broadcast Variables]({{ site.baseurl 
}}/dev/batch/index.html#broadcast-variables)), and for accessing runtime
-information such as accumulators and counters (see
-[Accumulators and Counters](#accumulators--counters)), and information
-on iterations (see [Iterations]({{ site.baseurl }}/dev/batch/iterations.html)).
+除了用户自定义的功能(map,reduce 等),Rich functions 
还提供了四个方法:`open`、`close`、`getRuntimeContext` 和
+`setRuntimeContext`。这些对于参数化功能很有用
+(参阅 [给函数传递参数]({{ site.baseurl 
}}/zh/dev/batch/index.html#passing-parameters-to-functions)),
+创建和最终确定本地状态,访问广播变量(参阅
+[广播变量]({{ site.baseurl 
}}/zh/dev/batch/index.html#broadcast-variables)),以及访问运行时信息,例如累加器和计数器(参阅
+[累加器和计数器](#累加器和计数器)),以及迭代器的相关信息(参阅 [迭代器]({{ site.baseurl 
}}/zh/dev/batch/iterations.html))。
 
 {% top %}
 
-## Accumulators & Counters
+
 
-Accumulators are simple constructs with an **add operation** and a **final 
accumulated result**,
-which is available after the job ended.
+## 累加器和计数器
 
-The most straightforward accumulator is a **counter**: You can increment it 
using the
-```Accumulator.add(V value)``` method. At the end of the job Flink will sum up 
(merge) all partial
-results and send the result to the client. Accumulators are useful during 
debugging or if you
-quickly want to find out more about your data.
+累加器是具有**加法运算**和**最终累加结果**的一种简单结构,可在作业结束后使用。
 
-Flink currently has the following **built-in accumulators**. Each of them 
implements the
-{% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Accumulator.java
 "Accumulator" %}
-interface.
+最简单的累加器就是**计数器**: 你可以使用
+```Accumulator.add(V value)``` 方法将其递增。在作业结束时,Flink 会汇总(合并)所有部分的结果并将其发送给客户端。
+在调试过程中或在你想快速了解有关数据更多信息时,累加器作用很大。
+
+Flink 目前有如下**内置累加器**。每个都实现了
+{% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Accumulator.java
 "累加器" %}
+接口。
 
 - {% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/IntCounter.java
 "__IntCounter__" %},
   {% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/LongCounter.java
 "__LongCounter__" %}
-  and {% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/DoubleCounter.java
 "__DoubleCounter__" %}:
-  See below for an example using a counter.
-- {% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
 "__Histogram__" %}:
-  A histogram implementation for a discrete number of bins. Internally it is 
just a map from Integer
-  to Integer. You can use this to compute distributions of values, e.g. the 
distribution of
-  words-per-line for a word count program.
+  和 {% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/DoubleCounter.java
 "__DoubleCounter__" %}:
+  有关使用计数器的示例,请参见下文。
+- {% gh_link 
/flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java
 "__直方图__" %}:
+  离散数量的柱状直方图实现。在内部,它只是整形到整形的映射。你可以使用它来计算值的分布,例如,单词计数程序的每行单词的分布情况。
 
-__How to use accumulators:__
+__如何使用累加器:__
 
-First you have to create an accumulator object (here a counter) in the 
user-defined transformation
-function where you want to use it.
+首先,你要在需要使用累加器的用户自定义的转换函数中创建一个累加器对象(此处是计数器)。

Review comment:
   ```suggestion
   首先,在需要使用累加器的用户自定义函数中创建一个累加器对象(此处是计数器)。
   ```

##
File path: docs/dev/user_defined_functions.zh.md
##
@@ -147,95 +153,77 @@ data.map (new RichMapFunction[String, Int] {
 
 
 
-Rich functions provide, in addition to the user-defined function (map,
-reduce, etc), four methods: `open`, `close`, `getRuntimeContext`, and
-`setRuntimeContext`. These are useful for parameterizing the function
-(see [Passing Parameters to Functions]({{ site.baseurl 
}}/dev/batch/index.html#passing-parameters-to-functions)),
-creating and finalizing local state, accessing broadcast variables (see
-[Broadcast Variables]({{ site.baseurl 
}}/dev/batch/index.html#broadcast-variables)), and for accessing runtime
-information such as accumulators and counters (see
-[Accumulators and Counters](#accumulators--counters)), and information
-on iterations (see [Iterations]({{ site.baseurl }}/dev/batch/iterations.html)).
+除了用户自定义的功能(map,reduce 等),Rich functions 
还提供了四个方法:`open`、`close`、`getRuntimeContext` 和
+`setRuntimeContext`。这些对于参数化功能很有用
+(参阅 [给函数传递参数]({{ site.baseurl 
}}/zh/dev/batch/index.html#passing-parameters-to-functi

[jira] [Updated] (FLINK-18695) Allow NettyBufferPool to allocate heap buffers

2020-08-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18695?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18695:
---
Labels: pull-request-available  (was: )

> Allow NettyBufferPool to allocate heap buffers
> --
>
> Key: FLINK-18695
> URL: https://issues.apache.org/jira/browse/FLINK-18695
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Chesnay Schepler
>Assignee: Yun Gao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> in 4.1.43 netty made a change to their SslHandler to always use heap buffers 
> for JDK SSLEngine implementations, to avoid an additional memory copy.
> However, our {{NettyBufferPool}} forbids heap buffer allocations.
> We will either have to allow heap buffer allocations, or create a custom 
> SslHandler implementation that does not use heap buffers (although this seems 
> ill-adviced?).
> /cc [~sewen] [~uce] [~NicoK] [~zjwang] [~pnowojski]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13109:
URL: https://github.com/apache/flink/pull/13109#issuecomment-671381707


   
   ## CI report:
   
   * 3d854c62355f9049062a7ae6a908dcceecd9c213 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5932)
 
   * 87ec25bce258c9fc953084701dd3acf7d96ac9e2 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5951)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] gaoyunhaii opened a new pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers

2020-08-27 Thread GitBox


gaoyunhaii opened a new pull request #13272:
URL: https://github.com/apache/flink/pull/13272


   ## What is the purpose of the change
   
   This PR modifies the `NettyBufferPool` to also allocate direct buffers for 
the heap buffer request. This enables us not to change the memory footprint 
when upgrade Netty to 4.1.50-FINAL. In the future we could further decide how 
to adjust the Netty memory management. 
   
   ## Brief change log
   
   - 3ef9f728f6a8005a7f91846bf32df96ee718c626 returns direct buffers for all 
the heap buffer allocation methods.
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): **no**
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: **no**
 - The serializers: **no**
 - The runtime per-record code paths (performance sensitive): **no**
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: **no**
 - The S3 file system connector: **no**
   
   ## Documentation
   
 - Does this pull request introduce a new feature? **no**
 - If yes, how is the feature documented? **not applicable**
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13109:
URL: https://github.com/apache/flink/pull/13109#issuecomment-671381707


   
   ## CI report:
   
   * 3d854c62355f9049062a7ae6a908dcceecd9c213 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5932)
 
   * 87ec25bce258c9fc953084701dd3acf7d96ac9e2 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] liming30 commented on a change in pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.

2020-08-27 Thread GitBox


liming30 commented on a change in pull request #13109:
URL: https://github.com/apache/flink/pull/13109#discussion_r478836972



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/collector/selector/DirectedOutputsCollector.java
##
@@ -0,0 +1,67 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.collector.selector;
+
+import org.apache.flink.streaming.api.operators.Output;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+
+import java.util.Map;
+
+/**
+ * The selected outputs collector will send records to the default output,
+ * and output matching outputNames.
+ *
+ * @param  The type of the elements that can be emitted.
+ */
+public class DirectedOutputsCollector implements 
SelectedOutputsCollector {
+
+   private final Output>[] selectAllOutputs;
+   private final Map>[]> outputMap;
+
+   public DirectedOutputsCollector(
+   Output>[] selectAllOutputs,
+   Map>[]> outputMap) {
+   this.selectAllOutputs = selectAllOutputs;
+   this.outputMap = outputMap;
+   }
+
+   @Override
+   public boolean collect(Iterable outputNames, StreamRecord 
record) {
+   boolean emitted = false;
+
+   if (selectAllOutputs.length > 0) {
+   collect(selectAllOutputs, record);
+   emitted = true;
+   }
+
+   for (String outputName : outputNames) {
+   Output>[] outputList = 
outputMap.get(outputName);
+   if (outputList != null && outputList.length > 0) {
+   collect(outputList, record);
+   emitted = true;
+   }
+   }

Review comment:
   In the old implementation via `set`, even if the same `output` appears 
multiple times in `outputNames`, it will only be sent once. Now it will send 
multiple times and I am not sure if this behavior is correct.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] liming30 commented on a change in pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.

2020-08-27 Thread GitBox


liming30 commented on a change in pull request #13109:
URL: https://github.com/apache/flink/pull/13109#discussion_r478431177



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/collector/selector/SelectedOutputsCollectorImpl.java
##
@@ -0,0 +1,61 @@
+package org.apache.flink.streaming.api.collector.selector;
+
+import org.apache.flink.streaming.api.operators.Output;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+
+import java.util.Map;
+
+/**
+ * The selected outputs collector will send records to the default output,
+ * and output matching outputNames.
+ *
+ * @param  The type of the elements that can be emitted.
+ */
+public class SelectedOutputsCollectorImpl implements 
SelectedOutputsCollector {
+
+   private final Output>[] selectAllOutputs;
+   private final Map>[]> outputMap;
+
+   private final boolean objectReuse;
+
+   public SelectedOutputsCollectorImpl(
+   Output>[] selectAllOutputs,
+   Map>[]> outputMap,
+   boolean objectReuse) {
+   this.selectAllOutputs = selectAllOutputs;
+   this.outputMap = outputMap;
+   this.objectReuse = objectReuse;
+   }
+
+   @Override
+   public boolean collect(Iterable outputNames, StreamRecord 
record) {
+   boolean emitted = false;
+
+   if (selectAllOutputs.length > 0) {
+   collect(selectAllOutputs, record);
+   emitted = true;
+   }
+
+   for (String outputName : outputNames) {
+   Output>[] outputList = 
outputMap.get(outputName);
+   if (outputList != null && outputList.length > 0) {
+   collect(outputList, record);
+   emitted = true;
+   }
+   }

Review comment:
   In the old implementation via `set`, even if the same `output` appears 
multiple times in `outputNames`, it will only be sent once. Now it will send 
multiple times and I am not sure if this behavior is correct.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] dianfu commented on pull request #13193: [FLINK-18918][python][docs] Add dedicated connector documentation for Python Table API

2020-08-27 Thread GitBox


dianfu commented on pull request #13193:
URL: https://github.com/apache/flink/pull/13193#issuecomment-682331400


   @hequn8128 Thanks for the update. LGTM.
   
   @sjwiesman @morsapaes could you take a further look at of the latest PR? 
Thanks a lot!



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13271:
URL: https://github.com/apache/flink/pull/13271#issuecomment-682323547


   
   ## CI report:
   
   * dcfea8ba4f522e9cd1a87022eadb10438902f22c Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5950)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread GitBox


flinkbot commented on pull request #13271:
URL: https://github.com/apache/flink/pull/13271#issuecomment-682323547


   
   ## CI report:
   
   * dcfea8ba4f522e9cd1a87022eadb10438902f22c UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19063) Support join late event from dimension table side in temporal table join

2020-08-27 Thread Paul Lin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186253#comment-17186253
 ] 

Paul Lin commented on FLINK-19063:
--

WRT event-time temporal table join, as an initial and naive thought, I think we 
can leverage the watermarks of both streams (may require some watermark 
mechanism changes). We should ensure the build side watermark is greater than 
the one of the probe size, if not, maybe we can keep the unjoined data of the 
probe side table before the build side watermark(plus allowed lateness) in the 
states.

When a new element of the build side shows up, it triggers the unjoined data in 
the states to re-join, produce the join result (if any), and be removed from 
the states. The process would be similar to event-time interval join.

> Support join late event from dimension table side in temporal table join 
> -
>
> Key: FLINK-19063
> URL: https://issues.apache.org/jira/browse/FLINK-19063
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Reporter: Leonard Xu
>Priority: Major
>
> To join late event from dimension table side in temporal table join is a 
> common user case
> from user-zh mail list[1][2].
> And another similar user case is how to enable the faster stream to wait the 
> slower stream  in regular stream join[3]. 
> I think we can discuss how to support these user cases.   
>  
>  
> [1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html]
> [2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585]
> [3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] RocMarshal commented on a change in pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread GitBox


RocMarshal commented on a change in pull request #13271:
URL: https://github.com/apache/flink/pull/13271#discussion_r478825509



##
File path: docs/monitoring/logging.zh.md
##
@@ -100,15 +106,13 @@ import org.slf4j.Logger
 Logger LOG = LoggerFactory.getLogger(Foobar.class)
 {% endhighlight %}
 
-In order to benefit most from slf4j, it is recommended to use its placeholder 
mechanism.
-Using placeholders allows to avoid unnecessary string constructions in case 
that the logging level is set so high that the message would not be logged.
-The syntax of placeholders is the following:
+为了最大限度地利用 slf4j,建议使用其占位符机制。使用占位符可以避免不必要的字符串构造,以防日志级别设置得太高而不会记录消息。占位符的语法如下:
 
 {% highlight java %}
 LOG.info("This message contains {} placeholders. {}", 2, "Yippie");
 {% endhighlight %}
 
-Placeholders can also be used in conjunction with exceptions which shall be 
logged.
+占位符也可以与应记录的异常一起使用。

Review comment:
   ```suggestion
   占位符也可以和要记录的异常一起使用。
   ```





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread GitBox


flinkbot commented on pull request #13271:
URL: https://github.com/apache/flink/pull/13271#issuecomment-682319220


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit a50c1399395cbc32008f02ffdb0380c897f522a4 (Fri Aug 28 
04:33:59 UTC 2020)
   
✅no warnings
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19043) Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-19043:
---
Labels: Documentation Translation pull-request-available translation-zh  
(was: Documentation Translation translation-zh)

> Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
> -
>
> Key: FLINK-19043
> URL: https://issues.apache.org/jira/browse/FLINK-19043
> Project: Flink
>  Issue Type: Improvement
>  Components: chinese-translation, Documentation
>Affects Versions: 1.10.0, 1.10.1, 1.11.0, 1.10.2, 1.11.1
>Reporter: Roc Marshal
>Assignee: Roc Marshal
>Priority: Major
>  Labels: Documentation, Translation, pull-request-available, 
> translation-zh
>
> The page url is : 
> [Logging|https://ci.apache.org/projects/flink/flink-docs-release-1.11/monitoring/logging.html]
>  The markdown file location is : flink/docs/monitoring/logging.zh.md



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] RocMarshal opened a new pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese

2020-08-27 Thread GitBox


RocMarshal opened a new pull request #13271:
URL: https://github.com/apache/flink/pull/13271


   
   
   ## What is the purpose of the change
   
   *Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese*
   
   
   ## Brief change log
   
   *Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese*
 - *The page url is : Logging 
https://ci.apache.org/projects/flink/flink-docs-release-1.11/monitoring/logging.html*
 - *The markdown file location is : flink/docs/monitoring/logging.zh.md*
   
   
   ## Verifying this change
   
   *Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese*
   
   A pure translation work in the `Documentation` module.
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable & docs
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13270:
URL: https://github.com/apache/flink/pull/13270#issuecomment-682310308


   
   ## CI report:
   
   * 04b974f56351d188b12bfe7c22a2fb8cb6e8f68e Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5948)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13249: [FLINK-14087][datastream] Clone the StreamPartitioner to avoid being shared at runtime.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13249:
URL: https://github.com/apache/flink/pull/13249#issuecomment-680775587


   
   ## CI report:
   
   * 4eae231a2cd3edbb2ba34f7bef983f64a4c4fefa Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5915)
 
   * 07e11ae0c700a4172a854ae4fd108b483bde6003 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5947)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec

2020-08-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-18900.

Resolution: Fixed

> HiveCatalog should error out when listing partitions with an invalid spec
> -
>
> Key: FLINK-18900
> URL: https://issues.apache.org/jira/browse/FLINK-18900
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1
>Reporter: Rui Li
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>
> Take the following case as an example:
> {code}
> create table tbl (x int) partitioned by (p int);
> alter table tbl add partition (p=1);
> {code}
> If we list partitions with partition spec {{foo=1}}, HiveCatalog returns 
> partition {{p=1}}, which is wrong.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec

2020-08-27 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186237#comment-17186237
 ] 

Jingsong Lee commented on FLINK-18900:
--

Revert interface modification for 1.11: a5767906916548ca51bf2d4b9e75c833ea6522a6

> HiveCatalog should error out when listing partitions with an invalid spec
> -
>
> Key: FLINK-18900
> URL: https://issues.apache.org/jira/browse/FLINK-18900
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1
>Reporter: Rui Li
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>
> Take the following case as an example:
> {code}
> create table tbl (x int) partitioned by (p int);
> alter table tbl add partition (p=1);
> {code}
> If we list partitions with partition spec {{foo=1}}, HiveCatalog returns 
> partition {{p=1}}, which is wrong.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi merged pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions

2020-08-27 Thread GitBox


JingsongLi merged pull request #13269:
URL: https://github.com/apache/flink/pull/13269


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices

2020-08-27 Thread GitBox


flinkbot commented on pull request #13270:
URL: https://github.com/apache/flink/pull/13270#issuecomment-682310308


   
   ## CI report:
   
   * 04b974f56351d188b12bfe7c22a2fb8cb6e8f68e UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13249: [FLINK-14087][datastream] Clone the StreamPartitioner to avoid being shared at runtime.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13249:
URL: https://github.com/apache/flink/pull/13249#issuecomment-680775587


   
   ## CI report:
   
   * 4eae231a2cd3edbb2ba34f7bef983f64a4c4fefa Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5915)
 
   * 07e11ae0c700a4172a854ae4fd108b483bde6003 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19081) Deprecate TemporalTableFunction and Table#createTemporalTableFunction()

2020-08-27 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19081?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu updated FLINK-19081:
---
Summary: Deprecate TemporalTableFunction and 
Table#createTemporalTableFunction()  (was: Deprecate TemporalTableFunction and 
Table$createTemporalTableFunction())

> Deprecate TemporalTableFunction and Table#createTemporalTableFunction()
> ---
>
> Key: FLINK-19081
> URL: https://issues.apache.org/jira/browse/FLINK-19081
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Leonard Xu
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19082) Add docs for temporal table and temporal table join

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19082:
--

 Summary: Add docs for temporal table and temporal table join
 Key: FLINK-19082
 URL: https://issues.apache.org/jira/browse/FLINK-19082
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19081) Deprecate TemporalTableFunction and Table$createTemporalTableFunction()

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19081:
--

 Summary: Deprecate TemporalTableFunction and 
Table$createTemporalTableFunction()
 Key: FLINK-19081
 URL: https://issues.apache.org/jira/browse/FLINK-19081
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19080) Materialize timeindicator data type in the right input of temporal join

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19080:
--

 Summary: Materialize timeindicator data type in the right input of 
temporal join
 Key: FLINK-19080
 URL: https://issues.apache.org/jira/browse/FLINK-19080
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19079) Support row time deduplicate operator

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19079:
--

 Summary: Support row time deduplicate operator
 Key: FLINK-19079
 URL: https://issues.apache.org/jira/browse/FLINK-19079
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19077) Improve process time temporal join operator

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19077:
--

 Summary: Improve process time temporal join operator
 Key: FLINK-19077
 URL: https://issues.apache.org/jira/browse/FLINK-19077
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19078) Import rowtime join temporal operator

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19078:
--

 Summary: Import rowtime join temporal operator
 Key: FLINK-19078
 URL: https://issues.apache.org/jira/browse/FLINK-19078
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19076) Import rule to deal Temporal Join condition

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19076:
--

 Summary: Import rule to deal Temporal Join condition
 Key: FLINK-19076
 URL: https://issues.apache.org/jira/browse/FLINK-19076
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19075) Infer changelog trait for temporal join node

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19075:
--

 Summary: Infer  changelog trait for temporal join node
 Key: FLINK-19075
 URL: https://issues.apache.org/jira/browse/FLINK-19075
 Project: Flink
  Issue Type: Bug
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19073) Improve streamExecTemporalJoinRule

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19073:
--

 Summary: Improve streamExecTemporalJoinRule
 Key: FLINK-19073
 URL: https://issues.apache.org/jira/browse/FLINK-19073
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19074) Materialize timeindicator in the right input of temporal join

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19074:
--

 Summary: Materialize timeindicator in the right input of temporal 
join
 Key: FLINK-19074
 URL: https://issues.apache.org/jira/browse/FLINK-19074
 Project: Flink
  Issue Type: Bug
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices

2020-08-27 Thread GitBox


flinkbot commented on pull request #13270:
URL: https://github.com/apache/flink/pull/13270#issuecomment-682306916


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 04b974f56351d188b12bfe7c22a2fb8cb6e8f68e (Fri Aug 28 
03:38:24 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-19072) Import Temporal Table join rule

2020-08-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-19072:
--

 Summary: Import Temporal Table join rule
 Key: FLINK-19072
 URL: https://issues.apache.org/jira/browse/FLINK-19072
 Project: Flink
  Issue Type: Sub-task
Reporter: Leonard Xu






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] sdlcwangsong opened a new pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices

2020-08-27 Thread GitBox


sdlcwangsong opened a new pull request #13270:
URL: https://github.com/apache/flink/pull/13270


   
   
   ## What is the purpose of the change
   
   *(For example: This pull request makes task deployment go through the blob 
server, rather than through RPC. That way we avoid re-transferring them on each 
deployment (during recovery).)*
   
   
   ## Brief change log
   
   *(for example:)*
 - *The TaskInfo is stored in the blob store on job creation time as a 
persistent artifact*
 - *Deployments RPC transmits only the blob storage reference*
 - *TaskManagers retrieve the TaskInfo from the blob cache*
   
   
   ## Verifying this change
   
   *(Please pick either of the following options)*
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   *(or)*
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   *(or)*
   
   This change added tests and can be verified as follows:
   
   *(example:)*
 - *Added integration tests for end-to-end deployment with large payloads 
(100MB)*
 - *Extended integration test for recovery after master (JobManager) 
failure*
 - *Added test that validates that TaskInfo is transferred only once across 
recoveries*
 - *Manually verified the change by running a 4 node cluser with 2 
JobManagers and 4 TaskManagers, a stateful streaming program, and killing one 
JobManager and two TaskManagers during the execution, verifying that recovery 
happens correctly.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented?
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13263:
URL: https://github.com/apache/flink/pull/13263#issuecomment-681892463


   
   ## CI report:
   
   * 990aa0f33e2fa93d40ea011082fedc6eeb37be44 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5922)
 
   * 4f0f42441fadc0961dbdecc9f1cf5ced660387d4 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5946)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-19070) Hive connector should throw a meaningful exception if user reads/writes ACID tables

2020-08-27 Thread Rui Li (Jira)
Rui Li created FLINK-19070:
--

 Summary: Hive connector should throw a meaningful exception if 
user reads/writes ACID tables
 Key: FLINK-19070
 URL: https://issues.apache.org/jira/browse/FLINK-19070
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Hive
Reporter: Rui Li
 Fix For: 1.12.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Nicholas Jiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186232#comment-17186232
 ] 

Nicholas Jiang edited comment on FLINK-19064 at 8/28/20, 3:22 AM:
--

[~ZhuShang] Please pay attention to the defintion of method openInputFormat, 
which javadoc mentions that Resources should be allocated in this method. (e.g. 
database connections, cache, etc.).


was (Author: nicholasjiang):
[~ZhuShang] Please pay attention to the defintion of method openInputFormat, 
which javadoc mentions that Resources should be allocated in this method. (e.g. 
database connections, cache, etc.).

 

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Nicholas Jiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186232#comment-17186232
 ] 

Nicholas Jiang commented on FLINK-19064:


[~ZhuShang] Please pay attention to the defintion of method openInputFormat, 
which javadoc mentions that Resources should be allocated in this method. (e.g. 
database connections, cache, etc.).

 

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19071) Some Hive window functions are not supported

2020-08-27 Thread Rui Li (Jira)
Rui Li created FLINK-19071:
--

 Summary: Some Hive window functions are not supported
 Key: FLINK-19071
 URL: https://issues.apache.org/jira/browse/FLINK-19071
 Project: Flink
  Issue Type: Task
  Components: Connectors / Hive
Reporter: Rui Li






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19069) finalizeOnMaster takes too much time and client timeouts

2020-08-27 Thread Kenneth William Krugler (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186231#comment-17186231
 ] 

Kenneth William Krugler commented on FLINK-19069:
-

I'd recently posted about a related issue to the dev mailing list, where I 
asked:

{quote}[...] the default behavior of Hadoop’s FileOutputCommitter (with 
algorithm == 1) is to put files in task-specific sub-dirs.

It’s depending on a post-completion “merge paths” action to be taken by what is 
(for Hadoop) the Application Master.

I assume that when running on a real cluster, the 
HadoopOutputFormat.finalizeGlobal() method’s call to commitJob() would do this, 
but it doesn’t seem to be happening when I run locally.

If I set the algorithm version to 2, then “merge paths” is handled by 
FileOutputCommitter immediately, and the HadoopOutputFormat code finds files in 
the expected location.

Wondering if Flink should always be using version 2 of the algorithm, as that’s 
more performant when there are a lot of results (which is why it was added).

{quote}

> finalizeOnMaster takes too much time and client timeouts
> 
>
> Key: FLINK-19069
> URL: https://issues.apache.org/jira/browse/FLINK-19069
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Task
>Affects Versions: 1.9.0
>Reporter: Jiayi Liao
>Priority: Major
>
> Currently we execute {{finalizeOnMaster}} in JM's main thread, which may 
> stuck the JM for a very long time and client timeouts eventually. 
> For example, we'd like to write data to HDFS  and commit files on JM, which 
> takes more than ten minutes to commit tens of thousands files.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13263:
URL: https://github.com/apache/flink/pull/13263#issuecomment-681892463


   
   ## CI report:
   
   * 990aa0f33e2fa93d40ea011082fedc6eeb37be44 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5922)
 
   * 4f0f42441fadc0961dbdecc9f1cf5ced660387d4 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-16824) FLIP-132 Temporal Table DDL and Temporal Table Join

2020-08-27 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu updated FLINK-16824:
---
Summary: FLIP-132 Temporal Table DDL and Temporal Table Join  (was: 
Creating Temporal Table Function via DDL)

> FLIP-132 Temporal Table DDL and Temporal Table Join
> ---
>
> Key: FLINK-16824
> URL: https://issues.apache.org/jira/browse/FLINK-16824
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Reporter: Konstantin Knauf
>Assignee: Leonard Xu
>Priority: Major
> Fix For: 1.12.0
>
>
> Currently, a Temporal Table Function can only be created via the Table API or 
> indirectly via the configuration file of the SQL Client. 
> It would be great, if this was also possible in pure SQL via a DDL statement. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19063) Support join late event from dimension table side in temporal table join

2020-08-27 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu updated FLINK-19063:
---
Description: 
To join late event from dimension table side in temporal table join is a common 
user case

from user-zh mail list[1][2].

And another similar user case is how to enable the faster stream to wait the 
slower stream  in regular stream join[3]. 

I think we can discuss how to support these user cases.   

 

 

[1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html]

[2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585]

[3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436]

  was:
To join late event from dimension table side in temporal table join is a common 
user case

from user-zh mail list[1][3].

And another similar user case is how to enable the faster stream to wait the 
slower stream  in regular stream join[3]. 

I think we can discuss how to support these user cases.   

 

 

[1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html]

[2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585]

[3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436]


> Support join late event from dimension table side in temporal table join 
> -
>
> Key: FLINK-19063
> URL: https://issues.apache.org/jira/browse/FLINK-19063
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Reporter: Leonard Xu
>Priority: Major
>
> To join late event from dimension table side in temporal table join is a 
> common user case
> from user-zh mail list[1][2].
> And another similar user case is how to enable the faster stream to wait the 
> slower stream  in regular stream join[3]. 
> I think we can discuss how to support these user cases.   
>  
>  
> [1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html]
> [2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585]
> [3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-19005) used metaspace grow on every execution

2020-08-27 Thread ShenDa (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186221#comment-17186221
 ] 

ShenDa edited comment on FLINK-19005 at 8/28/20, 2:57 AM:
--

[~chesnay] 
Thanks for your detailed instruction. 
But I still think there's maybe something wrong in Flink. I find that the 
JdbcInputFormat & JdbcOutputFormat is key reason cause the Metaspace OOM, 
because the java.sql.DriverManager doesn't release the reference of the Driver. 
The DriverManager is loaded by java.internal.ClassLoader but the driver is 
loaded by ChildFisrtClassLoader, which means the ChildFirstClassLoader can't be 
garbage collected according analyzation of dump file.  
The following code is used by me to reproduce the issue and  I use 
org.postgresql.Driver as jdbc Driver.
{code:java}
public static void main(String[] args) throws Exception {
EnvironmentSettings envSettings = EnvironmentSettings.newInstance()
.useBlinkPlanner() !origin-jdbc-inputformat.png! 
.inBatchMode()
.build();
TableEnvironment tEnv = TableEnvironment.create(envSettings);

tEnv.executeSql(
"CREATE TABLE " + INPUT_TABLE + "(" +
"id BIGINT," +
"timestamp6_col TIMESTAMP(6)," +
"timestamp9_col TIMESTAMP(6)," +
"time_col TIME," +
"real_col FLOAT," +
"decimal_col DECIMAL(10, 4)" +
") WITH (" +
"  'connector.type'='jdbc'," +
"  'connector.url'='" + DB_URL + "'," +
"  'connector.table'='" + INPUT_TABLE + 
"'," +
"  'connector.USERNAME'='" + USERNAME + 
"'," +
"  'connector.PASSWORD'='" + PASSWORD + 
"'" +
")"
);

TableResult tableResult = tEnv.executeSql("SELECT timestamp6_col, 
decimal_col FROM " + INPUT_TABLE);
tableResult.collect();
}
{code}
And below diagram shows the Metaspace usage constantly growing up, and finally 
TaskManager will be offline.
 !origin-jdbc-inputformat.png! 


Additional, I try to fix this issue by appending the following code to the 
function closeInputFormat() which can finally trigger garbage collect in 
Metaspace.

{code:java}
try{
final Enumeration drivers = DriverManager.getDrivers();
while (drivers.hasMoreElements()) {
DriverManager.deregisterDriver(drivers.nextElement());
}
} catch (SQLException se) {
LOG.info("Inputformat couldn't be closed - " + se.getMessage());
}
{code}
The following diagram shows the usage of Metaspace will be decreased.
 !modified-jdbc-inputformat.png! 
So, do you think it's a flink problem, and should we create a new issue to fix.


was (Author: dadashen):
[~chesnay] 
Thanks for your detailed instruction. 
But I still think there's maybe something wrong in Flink. I find that the 
JdbcInputFormat & JdbcOutputFormat is key reason cause the Metaspace OOM, 
because the java.sql.DriverManager doesn't release the reference of the Driver. 
The DriverManager is loaded by java.internal.ClassLoader but the driver is 
loaded by ChildFisrtClassLoader, which means the ChildFirstClassLoader can't be 
garbage collected according analyzation of dump file.  
The following code is used by me to reproduce the issue and  I use 
org.postgresql.Driver as jdbc Driver.
{code:java}
public static void main(String[] args) throws Exception {
EnvironmentSettings envSettings = EnvironmentSettings.newInstance()
.useBlinkPlanner() !origin-jdbc-inputformat.png! 
.inBatchMode()
.build();
TableEnvironment tEnv = TableEnvironment.create(envSettings);

tEnv.executeSql(
"CREATE TABLE " + INPUT_TABLE + "(" +
"id BIGINT," +
"timestamp6_col TIMESTAMP(6)," +
"timestamp9_col TIMESTAMP(6)," +
"time_col TIME," +
"real_col FLOAT," +
"decimal_col DECIMAL(10, 4)" +
") WITH (" +
"  'connector.type'='jdbc'," +
"  'connector.url'='" + DB_URL + "'," +
"  'connector.table'='" + INPUT_TABLE + 
"'," +
  

[GitHub] [flink] xintongsong commented on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.

2020-08-27 Thread GitBox


xintongsong commented on pull request #13263:
URL: https://github.com/apache/flink/pull/13263#issuecomment-682296119


   Thanks for the review, @tillrohrmann 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Nicholas Jiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186224#comment-17186224
 ] 

Nicholas Jiang commented on FLINK-19064:


[~jark] I guess the meaning of [~ZhuShang] is that the javadoc of the method 
configure() is Creates a \{@link Scan} object and opens the \{@link HTable} 
connection, which means that HBase connection is defined to open in configure().

 

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Nicholas Jiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186224#comment-17186224
 ] 

Nicholas Jiang edited comment on FLINK-19064 at 8/28/20, 2:51 AM:
--

[~jark] I guess the meaning of [~ZhuShang] is that the javadoc of the method 
configure() is Creates a \{@link Scan} object and opens the \{@link HTable} 
connection, which means that HBase connection is defined to open in configure().


was (Author: nicholasjiang):
[~jark] I guess the meaning of [~ZhuShang] is that the javadoc of the method 
configure() is Creates a \{@link Scan} object and opens the \{@link HTable} 
connection, which means that HBase connection is defined to open in configure().

 

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19005) used metaspace grow on every execution

2020-08-27 Thread ShenDa (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186221#comment-17186221
 ] 

ShenDa commented on FLINK-19005:


[~chesnay] 
Thanks for your detailed instruction. 
But I still think there's maybe something wrong in Flink. I find that the 
JdbcInputFormat & JdbcOutputFormat is key reason cause the Metaspace OOM, 
because the java.sql.DriverManager doesn't release the reference of the Driver. 
The DriverManager is loaded by java.internal.ClassLoader but the driver is 
loaded by ChildFisrtClassLoader, which means the ChildFirstClassLoader can't be 
garbage collected according analyzation of dump file.  
The following code is used by me to reproduce the issue and  I use 
org.postgresql.Driver as jdbc Driver.
{code:java}
public static void main(String[] args) throws Exception {
EnvironmentSettings envSettings = EnvironmentSettings.newInstance()
.useBlinkPlanner() !origin-jdbc-inputformat.png! 
.inBatchMode()
.build();
TableEnvironment tEnv = TableEnvironment.create(envSettings);

tEnv.executeSql(
"CREATE TABLE " + INPUT_TABLE + "(" +
"id BIGINT," +
"timestamp6_col TIMESTAMP(6)," +
"timestamp9_col TIMESTAMP(6)," +
"time_col TIME," +
"real_col FLOAT," +
"decimal_col DECIMAL(10, 4)" +
") WITH (" +
"  'connector.type'='jdbc'," +
"  'connector.url'='" + DB_URL + "'," +
"  'connector.table'='" + INPUT_TABLE + 
"'," +
"  'connector.USERNAME'='" + USERNAME + 
"'," +
"  'connector.PASSWORD'='" + PASSWORD + 
"'" +
")"
);

TableResult tableResult = tEnv.executeSql("SELECT timestamp6_col, 
decimal_col FROM " + INPUT_TABLE);
tableResult.collect();
}
{code}
And below diagram shows the Metaspace usage constantly growing up, and finally 
TaskManager will be offline.
 !origin-jdbc-inputformat.png! 


Additional, I try to fix this issue by appending the following code to the 
function closeInputFormat() which can finally trigger garbage collect in 
Metaspace.

{code:java}
try{
final Enumeration drivers = DriverManager.getDrivers();
while (drivers.hasMoreElements()) {
DriverManager.deregisterDriver(drivers.nextElement());
}
} catch (SQLException se) {
LOG.info("Inputformat couldn't be closed - " + se.getMessage());
}
{code}
The following diagram shows the usage of Metaspace will be decreased.
 !modified-jdbc-inputformat.png! 

> used metaspace grow on every execution
> --
>
> Key: FLINK-19005
> URL: https://issues.apache.org/jira/browse/FLINK-19005
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, Runtime / Configuration, 
> Runtime / Coordination
>Affects Versions: 1.11.1
>Reporter: Guillermo Sánchez
>Assignee: Chesnay Schepler
>Priority: Major
> Attachments: heap_dump_after_10_executions.zip, 
> heap_dump_after_1_execution.zip, heap_dump_echo_lee.tar.xz, 
> modified-jdbc-inputformat.png, origin-jdbc-inputformat.png
>
>
> Hi !
> Im running a 1.11.1 flink cluster, where I execute batch jobs made with 
> DataSet API.
> I submit these jobs every day to calculate daily data.
> In every execution, cluster's used metaspace increase by 7MB and its never 
> released.
> This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i 
> need to restart the cluster to clean the metaspace
> taskmanager.memory.jvm-metaspace.size is set to 512mb
> Any idea of what could be causing this metaspace grow and why is it not 
> released ?
>  
> 
> === Summary ==
> 
> Case 1, reported by [~gestevez]:
> * Flink 1.11.1
> * Java 11
> * Maximum Metaspace size set to 512mb
> * Custom Batch job, submitted daily
> * Requires restart every 15 days after an OOM
>  Case 2, reported by [~Echo Lee]:
> * Flink 1.11.0
> * Java 11
> * G1GC
> * WordCount Batch job, submitted every second / every 5 minutes
> * eventually fails TaskExecutor with OOM
> Case 3, reported by [~DaDaShen]
> * Flink 1.11.0
> * Java 11
> * WordCount Batch job, submitted every 5 seconds
> * growing Metaspace, eventually OOM
>  



--
This message was se

[jira] [Updated] (FLINK-19005) used metaspace grow on every execution

2020-08-27 Thread ShenDa (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ShenDa updated FLINK-19005:
---
Attachment: modified-jdbc-inputformat.png

> used metaspace grow on every execution
> --
>
> Key: FLINK-19005
> URL: https://issues.apache.org/jira/browse/FLINK-19005
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, Runtime / Configuration, 
> Runtime / Coordination
>Affects Versions: 1.11.1
>Reporter: Guillermo Sánchez
>Assignee: Chesnay Schepler
>Priority: Major
> Attachments: heap_dump_after_10_executions.zip, 
> heap_dump_after_1_execution.zip, heap_dump_echo_lee.tar.xz, 
> modified-jdbc-inputformat.png, origin-jdbc-inputformat.png
>
>
> Hi !
> Im running a 1.11.1 flink cluster, where I execute batch jobs made with 
> DataSet API.
> I submit these jobs every day to calculate daily data.
> In every execution, cluster's used metaspace increase by 7MB and its never 
> released.
> This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i 
> need to restart the cluster to clean the metaspace
> taskmanager.memory.jvm-metaspace.size is set to 512mb
> Any idea of what could be causing this metaspace grow and why is it not 
> released ?
>  
> 
> === Summary ==
> 
> Case 1, reported by [~gestevez]:
> * Flink 1.11.1
> * Java 11
> * Maximum Metaspace size set to 512mb
> * Custom Batch job, submitted daily
> * Requires restart every 15 days after an OOM
>  Case 2, reported by [~Echo Lee]:
> * Flink 1.11.0
> * Java 11
> * G1GC
> * WordCount Batch job, submitted every second / every 5 minutes
> * eventually fails TaskExecutor with OOM
> Case 3, reported by [~DaDaShen]
> * Flink 1.11.0
> * Java 11
> * WordCount Batch job, submitted every 5 seconds
> * growing Metaspace, eventually OOM
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19069) finalizeOnMaster takes too much time and client timeouts

2020-08-27 Thread Jiayi Liao (Jira)
Jiayi Liao created FLINK-19069:
--

 Summary: finalizeOnMaster takes too much time and client timeouts
 Key: FLINK-19069
 URL: https://issues.apache.org/jira/browse/FLINK-19069
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Task
Affects Versions: 1.9.0
Reporter: Jiayi Liao


Currently we execute {{finalizeOnMaster}} in JM's main thread, which may stuck 
the JM for a very long time and client timeouts eventually. 

For example, we'd like to write data to HDFS  and commit files on JM, which 
takes more than ten minutes to commit tens of thousands files.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13269:
URL: https://github.com/apache/flink/pull/13269#issuecomment-682288332


   
   ## CI report:
   
   * c967e9837780e4ac82c668845bc35d1a38c47ee8 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5945)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186215#comment-17186215
 ] 

Jark Wu commented on FLINK-19064:
-

Hi [~ZhuShang], why the connection should be created in {{configure()}}?

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-19064) HBaseRowDataInputFormat is leaking resources

2020-08-27 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-19064:
---

Assignee: Nicholas Jiang

> HBaseRowDataInputFormat is leaking resources
> 
>
> Key: FLINK-19064
> URL: https://issues.apache.org/jira/browse/FLINK-19064
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Nicholas Jiang
>Priority: Critical
>
> {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, 
> which creates a connection to HBase that is not closed again.
> A user reported this problem on the user@ list: 
> https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19005) used metaspace grow on every execution

2020-08-27 Thread ShenDa (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ShenDa updated FLINK-19005:
---
Attachment: origin-jdbc-inputformat.png

> used metaspace grow on every execution
> --
>
> Key: FLINK-19005
> URL: https://issues.apache.org/jira/browse/FLINK-19005
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, Runtime / Configuration, 
> Runtime / Coordination
>Affects Versions: 1.11.1
>Reporter: Guillermo Sánchez
>Assignee: Chesnay Schepler
>Priority: Major
> Attachments: heap_dump_after_10_executions.zip, 
> heap_dump_after_1_execution.zip, heap_dump_echo_lee.tar.xz, 
> origin-jdbc-inputformat.png
>
>
> Hi !
> Im running a 1.11.1 flink cluster, where I execute batch jobs made with 
> DataSet API.
> I submit these jobs every day to calculate daily data.
> In every execution, cluster's used metaspace increase by 7MB and its never 
> released.
> This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i 
> need to restart the cluster to clean the metaspace
> taskmanager.memory.jvm-metaspace.size is set to 512mb
> Any idea of what could be causing this metaspace grow and why is it not 
> released ?
>  
> 
> === Summary ==
> 
> Case 1, reported by [~gestevez]:
> * Flink 1.11.1
> * Java 11
> * Maximum Metaspace size set to 512mb
> * Custom Batch job, submitted daily
> * Requires restart every 15 days after an OOM
>  Case 2, reported by [~Echo Lee]:
> * Flink 1.11.0
> * Java 11
> * G1GC
> * WordCount Batch job, submitted every second / every 5 minutes
> * eventually fails TaskExecutor with OOM
> Case 3, reported by [~DaDaShen]
> * Flink 1.11.0
> * Java 11
> * WordCount Batch job, submitted every 5 seconds
> * growing Metaspace, eventually OOM
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19068) Filter verbose pod events for KubernetesResourceManagerDriver

2020-08-27 Thread Xintong Song (Jira)
Xintong Song created FLINK-19068:


 Summary: Filter verbose pod events for 
KubernetesResourceManagerDriver
 Key: FLINK-19068
 URL: https://issues.apache.org/jira/browse/FLINK-19068
 Project: Flink
  Issue Type: Improvement
  Components: Deployment / Kubernetes
Reporter: Xintong Song


A status of a Kubernetes pod consists of many detailed fields. Currently, Flink 
receives pod {{MODIFIED}} events from the {{KubernetesPodsWatcher}} on every 
single change to these fields, many of which Flink does not care.

The verbose events will not affect the functionality of Flink, but will pollute 
the logs with repeated messages, because Flink only looks into the fields it 
interested in and those fields are identical.

E.g., when a task manager is stopped due to idle timeout, Flink receives 3 
events:
* MODIFIED: container terminated
* MODIFIED: {{deletionGracePeriodSeconds}} changes from 30 to 0, which is a 
Kubernetes internal status change after containers are gracefully terminated
* DELETED: Flink removes metadata of the terminated pod

Among the 3 messages, Flink is only interested in the 1st MODIFIED message, but 
will try to process all of them because the container status is terminated.

I propose to Filter the verbose events in 
{{KubernetesResourceManagerDriver.PodCallbackHandlerImpl}}, to only process the 
status changes interested by Flink. This probably requires recording the status 
of all living pods, to compare with the incoming events for detecting status 
changes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions

2020-08-27 Thread GitBox


flinkbot commented on pull request #13269:
URL: https://github.com/apache/flink/pull/13269#issuecomment-682288332


   
   ## CI report:
   
   * c967e9837780e4ac82c668845bc35d1a38c47ee8 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error

2020-08-27 Thread GitBox


flinkbot edited a comment on pull request #13268:
URL: https://github.com/apache/flink/pull/13268#issuecomment-682283788


   
   ## CI report:
   
   * 5357812df45c42860e95630891367a410f6bc34b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5944)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-15719) Exceptions when using scala types directly with the State Process API

2020-08-27 Thread Ying Z (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186212#comment-17186212
 ] 

Ying Z commented on FLINK-15719:


I pull a request here, [https://github.com/apache/flink/pull/13266]

could someone do a review, thanks.

> Exceptions when using scala types directly with the State Process API
> -
>
> Key: FLINK-15719
> URL: https://issues.apache.org/jira/browse/FLINK-15719
> Project: Flink
>  Issue Type: Bug
>  Components: API / State Processor
>Affects Versions: 1.9.1
>Reporter: Ying Z
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Major
>  Labels: pull-request-available
>
> I followed these steps to generate and read states:
>  # implements the example[1] `CountWindowAverage` in Scala(exactly same), and 
> run jobA => that makes good.
>  # execute `flink cancel -s ${JobID}` => savepoints was generated as expected.
>  # implements the example[2] `StatefulFunctionWithTime` in Scala(code below), 
> and run jobB => failed, exceptions shows that "Caused by: 
> org.apache.flink.util.StateMigrationException: The new key serializer must be 
> compatible."
> ReaderFunction code as below:
> {code:java}
> // code placeholder
>   class ReaderFunction extends KeyedStateReaderFunction[Long, (Long, Long)] {
> var countState: ValueState[(Long, Long)] = _
> override def open(parameters: Configuration): Unit = {
>   val stateDescriptor = new ValueStateDescriptor("average", 
> createTypeInformation[(Long, Long)])
>   countState = getRuntimeContext().getState(stateDescriptor)
> }override def readKey(key: Long, ctx: 
> KeyedStateReaderFunction.Context, out: Collector[(Long, Long)]): Unit = {
>   out.collect(countState.value())
> }
>   }
> {code}
> 1: 
> [https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/stream/state/state.html#using-managed-keyed-state]
>  
> 2: 
> [https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/libs/state_processor_api.html#keyed-state]
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-19060) Checkpoint not triggered when use broadcast stream

2020-08-27 Thread henvealf (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19060?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

henvealf closed FLINK-19060.

Resolution: Not A Bug

> Checkpoint not triggered when use broadcast stream
> --
>
> Key: FLINK-19060
> URL: https://issues.apache.org/jira/browse/FLINK-19060
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.11.1
>Reporter: henvealf
>Priority: Major
> Attachments: image-2020-08-27-16-41-23-699.png, 
> image-2020-08-27-16-44-37-442.png, image-2020-08-27-16-45-28-134.png, 
> image-2020-08-27-16-51-10-512.png
>
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> Code:
> !image-2020-08-27-16-51-10-512.png!
> KafkaSourceConfig:
> consumer.setStartFromGroupOffsets()
> Web UI:
>     !image-2020-08-27-16-45-28-134.png!
> Checkpoint always doesn't happen. Did I write something wrong?
> Thanks!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions

2020-08-27 Thread GitBox


flinkbot commented on pull request #13269:
URL: https://github.com/apache/flink/pull/13269#issuecomment-682286516


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit c967e9837780e4ac82c668845bc35d1a38c47ee8 (Fri Aug 28 
02:17:35 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] JingsongLi opened a new pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions

2020-08-27 Thread GitBox


JingsongLi opened a new pull request #13269:
URL: https://github.com/apache/flink/pull/13269


   Revert the modification of Catalog.listPartitions for version Compatibility.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec

2020-08-27 Thread Rui Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186205#comment-17186205
 ] 

Rui Li commented on FLINK-18900:


Perhaps we can revert the change in 1.11, given that 1.11 doesn't support SHOW 
PARTITIONS.

> HiveCatalog should error out when listing partitions with an invalid spec
> -
>
> Key: FLINK-18900
> URL: https://issues.apache.org/jira/browse/FLINK-18900
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1
>Reporter: Rui Li
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>
> Take the following case as an example:
> {code}
> create table tbl (x int) partitioned by (p int);
> alter table tbl add partition (p=1);
> {code}
> If we list partitions with partition spec {{foo=1}}, HiveCatalog returns 
> partition {{p=1}}, which is wrong.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec

2020-08-27 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186204#comment-17186204
 ] 

Jingsong Lee edited comment on FLINK-18900 at 8/28/20, 2:08 AM:


We should not modify method throws exception for {{Catalog.listPartitions}} in 
release-1.11.


was (Author: lzljs3620320):
We should not modify method throws exception for {{Catalog.listPartitions}}.

> HiveCatalog should error out when listing partitions with an invalid spec
> -
>
> Key: FLINK-18900
> URL: https://issues.apache.org/jira/browse/FLINK-18900
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1
>Reporter: Rui Li
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>
> Take the following case as an example:
> {code}
> create table tbl (x int) partitioned by (p int);
> alter table tbl add partition (p=1);
> {code}
> If we list partitions with partition spec {{foo=1}}, HiveCatalog returns 
> partition {{p=1}}, which is wrong.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec

2020-08-27 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186204#comment-17186204
 ] 

Jingsong Lee commented on FLINK-18900:
--

We should not modify method throws exception for {{Catalog.listPartitions}}.

> HiveCatalog should error out when listing partitions with an invalid spec
> -
>
> Key: FLINK-18900
> URL: https://issues.apache.org/jira/browse/FLINK-18900
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1
>Reporter: Rui Li
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>
> Take the following case as an example:
> {code}
> create table tbl (x int) partitioned by (p int);
> alter table tbl add partition (p=1);
> {code}
> If we list partitions with partition spec {{foo=1}}, HiveCatalog returns 
> partition {{p=1}}, which is wrong.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Reopened] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec

2020-08-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee reopened FLINK-18900:
--

> HiveCatalog should error out when listing partitions with an invalid spec
> -
>
> Key: FLINK-18900
> URL: https://issues.apache.org/jira/browse/FLINK-18900
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.11.1
>Reporter: Rui Li
>Assignee: Nicholas Jiang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.2
>
>
> Take the following case as an example:
> {code}
> create table tbl (x int) partitioned by (p int);
> alter table tbl add partition (p=1);
> {code}
> If we list partitions with partition spec {{foo=1}}, HiveCatalog returns 
> partition {{p=1}}, which is wrong.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error

2020-08-27 Thread GitBox


flinkbot commented on pull request #13268:
URL: https://github.com/apache/flink/pull/13268#issuecomment-682283788


   
   ## CI report:
   
   * 5357812df45c42860e95630891367a410f6bc34b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-19067) FileNotFoundException when run flink examples on standby JobManager

2020-08-27 Thread JieFang.He (Jira)
JieFang.He created FLINK-19067:
--

 Summary: FileNotFoundException when run flink examples on standby 
JobManager
 Key: FLINK-19067
 URL: https://issues.apache.org/jira/browse/FLINK-19067
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.11.1
Reporter: JieFang.He


1、When run examples/batch/WordCount.jar on standby JobManager,it will fail with 
the exception:

Caused by: java.io.FileNotFoundException: 
/data2/storageDir/default/blob/job_d29414828f614d5466e239be4d3889ac/blob_p-a2ebe1c5aa160595f214b4bd0f39d80e42ee2e93-f458f1c12dc023e78d25f191de1d7c4b
 (No such file or directory)
 at java.io.FileInputStream.open0(Native Method)
 at java.io.FileInputStream.open(FileInputStream.java:195)
 at java.io.FileInputStream.(FileInputStream.java:138)
 at 
org.apache.flink.core.fs.local.LocalDataInputStream.(LocalDataInputStream.java:50)
 at 
org.apache.flink.core.fs.local.LocalFileSystem.open(LocalFileSystem.java:143)
 at 
org.apache.flink.runtime.blob.FileSystemBlobStore.get(FileSystemBlobStore.java:105)
 at 
org.apache.flink.runtime.blob.FileSystemBlobStore.get(FileSystemBlobStore.java:87)
 at 
org.apache.flink.runtime.blob.BlobServer.getFileInternal(BlobServer.java:501)
 at 
org.apache.flink.runtime.blob.BlobServerConnection.get(BlobServerConnection.java:231)
 at 
org.apache.flink.runtime.blob.BlobServerConnection.run(BlobServerConnection.java:117)

 

2、Run examples success on other nodes

3、After run success on the other node, it can run success on the Standby 
JobManager. But run again will fail

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error

2020-08-27 Thread GitBox


flinkbot commented on pull request #13268:
URL: https://github.com/apache/flink/pull/13268#issuecomment-682280167


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 5357812df45c42860e95630891367a410f6bc34b (Fri Aug 28 
01:53:11 UTC 2020)
   
   **Warnings:**
* Documentation files were touched, but no `.zh.md` files: Update Chinese 
documentation or file Jira ticket.
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




  1   2   3   4   >