[jira] [Created] (FLINK-19083) Remove deprecated DataStream#split
Dawid Wysakowicz created FLINK-19083: Summary: Remove deprecated DataStream#split Key: FLINK-19083 URL: https://issues.apache.org/jira/browse/FLINK-19083 Project: Flink Issue Type: Sub-task Components: API / DataStream Reporter: Dawid Wysakowicz -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-19035) Remove deprecated DataStream#fold() method and all related classes
[ https://issues.apache.org/jira/browse/FLINK-19035?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dawid Wysakowicz updated FLINK-19035: - Fix Version/s: 2.0.0 > Remove deprecated DataStream#fold() method and all related classes > -- > > Key: FLINK-19035 > URL: https://issues.apache.org/jira/browse/FLINK-19035 > Project: Flink > Issue Type: Sub-task > Components: API / DataStream >Reporter: Dawid Wysakowicz >Priority: Major > Fix For: 2.0.0 > > > We should also agree if we want to remove those in 1.x version already or do > we want to wait for 2.0. > We should remove DataStream#fold and all related classes and methods such as > FoldFunction, FoldingState, FoldingStateDescriptor ... -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-19083) Remove deprecated DataStream#split
[ https://issues.apache.org/jira/browse/FLINK-19083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dawid Wysakowicz updated FLINK-19083: - Fix Version/s: 2.0.0 > Remove deprecated DataStream#split > -- > > Key: FLINK-19083 > URL: https://issues.apache.org/jira/browse/FLINK-19083 > Project: Flink > Issue Type: Sub-task > Components: API / DataStream >Reporter: Dawid Wysakowicz >Priority: Major > Fix For: 2.0.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.
flinkbot edited a comment on pull request #13263: URL: https://github.com/apache/flink/pull/13263#issuecomment-681892463 ## CI report: * 4f0f42441fadc0961dbdecc9f1cf5ced660387d4 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5946) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.
flinkbot edited a comment on pull request #13273: URL: https://github.com/apache/flink/pull/13273#issuecomment-682353427 ## CI report: * 0134fa06171742eb8ae840b2aced9530232bdb04 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5955) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13274: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…
flinkbot commented on pull request #13274: URL: https://github.com/apache/flink/pull/13274#issuecomment-682362109 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 7835ad0187076a7dae2ee7e8806337f1b3db01c1 (Fri Aug 28 06:55:05 UTC 2020) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lirui-apache opened a new pull request #13274: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…
lirui-apache opened a new pull request #13274: URL: https://github.com/apache/flink/pull/13274 …f partition value contains special characters ## What is the purpose of the change Fix the issue that HiveCatalog fails to get partition column stats if partition value contains special characters ## Brief change log - Escape partition name when get/alter partition column stats - Update test to cover the use case ## Verifying this change Updated test case ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: no - The serializers: no - The runtime per-record code paths (performance sensitive): no - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? no - If yes, how is the feature documented? NA This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] XComp commented on a change in pull request #13251: [FLINK-14435] Added memory configuration to TaskManagers REST endpoint
XComp commented on a change in pull request #13251: URL: https://github.com/apache/flink/pull/13251#discussion_r478866388 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/taskexecutor/TaskExecutorMemoryConfiguration.java ## @@ -0,0 +1,333 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.runtime.taskexecutor; + +import org.apache.flink.configuration.ConfigOption; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.configuration.MemorySize; + +import org.apache.flink.shaded.guava18.com.google.common.base.MoreObjects; +import org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonCreator; +import org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonInclude; +import org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonProperty; + +import java.io.Serializable; +import java.util.Objects; + +import static org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_OFF_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_METASPACE; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_FRACTION; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MAX; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MIN; +import static org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_FRACTION; +import static org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_SIZE; +import static org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_FRACTION; +import static org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MAX; +import static org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MIN; +import static org.apache.flink.configuration.TaskManagerOptions.TASK_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.TASK_OFF_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.TOTAL_PROCESS_MEMORY; + +/** + * TaskExecutorConfiguration collects the configuration of a TaskExecutor instance. + */ +public class TaskExecutorMemoryConfiguration implements Serializable { + + public static final String FIELD_NAME_FRAMEWORK_HEAP = "frameworkHeap"; + public static final String FIELD_NAME_TASK_HEAP = "taskHeap"; + + public static final String FIELD_NAME_FRAMEWORK_OFFHEAP = "frameworkOffHeap"; + public static final String FIELD_NAME_TASK_OFFHEAP = "taskOffHeap"; + + public static final String FIELD_NAME_NETWORK_MIN = "networkMin"; + public static final String FIELD_NAME_NETWORK_MAX = "networkMax"; + public static final String FIELD_NAME_NETWORK_FRACTION = "networkFraction"; + + public static final String FIELD_NAME_MANAGED_TOTAL = "managedTotal"; + public static final String FIELD_NAME_MANAGED_FRACTION = "managedFraction"; + + public static final String FIELD_NAME_METASPACE_MAX = "metaspaceMax"; + + public static final String FIELD_NAME_OVERHEAD_MIN = "overheadMin"; + public static final String FIELD_NAME_OVERHEAD_MAX = "overheadMax"; + public static final String FIELD_NAME_OVERHEAD_FRACTION = "overheadFraction"; + + public static final String FIELD_NAME_MEMORY_TOTAL = "memoryTotal"; + + @JsonProperty(FIELD_NAME_FRAMEWORK_HEAP) + @JsonInclude + private final Long frameworkHeap; + + @JsonProperty(FIELD_NAME_TASK_HEAP) + private final Long taskHeap; + + @JsonProperty(FIELD_NAME_FRAMEWORK_OFFHEAP) + private final Long frameworkOffHeap; + + @JsonProperty(FIELD_NAME_TASK_OFFHEAP) + private final Long taskOffHeap; + + @JsonProperty(FIELD_NAME_NETWORK_MIN) + private final Long networkMemoryMin; + @JsonProperty(FIELD_NAME_NETWORK_MAX) + private final Long networkMemoryMax; + @JsonProperty(FIELD_NAME_NETWORK_FRACTION) + private final Float networkMemoryFraction; + + @JsonProperty(FIELD_NAME_MANAGED_TOTAL) + private final Long managedMemoryTotal; +
[GitHub] [flink] XComp commented on a change in pull request #13251: [FLINK-14435] Added memory configuration to TaskManagers REST endpoint
XComp commented on a change in pull request #13251: URL: https://github.com/apache/flink/pull/13251#discussion_r478866224 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/taskexecutor/TaskExecutorMemoryConfiguration.java ## @@ -0,0 +1,333 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.runtime.taskexecutor; + +import org.apache.flink.configuration.ConfigOption; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.configuration.MemorySize; + +import org.apache.flink.shaded.guava18.com.google.common.base.MoreObjects; +import org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonCreator; +import org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonInclude; +import org.apache.flink.shaded.jackson2.com.fasterxml.jackson.annotation.JsonProperty; + +import java.io.Serializable; +import java.util.Objects; + +import static org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.FRAMEWORK_OFF_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_METASPACE; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_FRACTION; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MAX; +import static org.apache.flink.configuration.TaskManagerOptions.JVM_OVERHEAD_MIN; +import static org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_FRACTION; +import static org.apache.flink.configuration.TaskManagerOptions.MANAGED_MEMORY_SIZE; +import static org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_FRACTION; +import static org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MAX; +import static org.apache.flink.configuration.TaskManagerOptions.NETWORK_MEMORY_MIN; +import static org.apache.flink.configuration.TaskManagerOptions.TASK_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.TASK_OFF_HEAP_MEMORY; +import static org.apache.flink.configuration.TaskManagerOptions.TOTAL_PROCESS_MEMORY; + +/** + * TaskExecutorConfiguration collects the configuration of a TaskExecutor instance. + */ +public class TaskExecutorMemoryConfiguration implements Serializable { + + public static final String FIELD_NAME_FRAMEWORK_HEAP = "frameworkHeap"; + public static final String FIELD_NAME_TASK_HEAP = "taskHeap"; + + public static final String FIELD_NAME_FRAMEWORK_OFFHEAP = "frameworkOffHeap"; + public static final String FIELD_NAME_TASK_OFFHEAP = "taskOffHeap"; + + public static final String FIELD_NAME_NETWORK_MIN = "networkMin"; + public static final String FIELD_NAME_NETWORK_MAX = "networkMax"; + public static final String FIELD_NAME_NETWORK_FRACTION = "networkFraction"; + + public static final String FIELD_NAME_MANAGED_TOTAL = "managedTotal"; + public static final String FIELD_NAME_MANAGED_FRACTION = "managedFraction"; + + public static final String FIELD_NAME_METASPACE_MAX = "metaspaceMax"; + + public static final String FIELD_NAME_OVERHEAD_MIN = "overheadMin"; + public static final String FIELD_NAME_OVERHEAD_MAX = "overheadMax"; + public static final String FIELD_NAME_OVERHEAD_FRACTION = "overheadFraction"; + + public static final String FIELD_NAME_MEMORY_TOTAL = "memoryTotal"; + + @JsonProperty(FIELD_NAME_FRAMEWORK_HEAP) + @JsonInclude + private final Long frameworkHeap; + + @JsonProperty(FIELD_NAME_TASK_HEAP) + private final Long taskHeap; + + @JsonProperty(FIELD_NAME_FRAMEWORK_OFFHEAP) + private final Long frameworkOffHeap; + + @JsonProperty(FIELD_NAME_TASK_OFFHEAP) + private final Long taskOffHeap; + + @JsonProperty(FIELD_NAME_NETWORK_MIN) + private final Long networkMemoryMin; + @JsonProperty(FIELD_NAME_NETWORK_MAX) + private final Long networkMemoryMax; + @JsonProperty(FIELD_NAME_NETWORK_FRACTION) + private final Float networkMemoryFraction; + + @JsonProperty(FIELD_NAME_MANAGED_TOTAL) + private final Long managedMemoryTotal; +
[GitHub] [flink] tillrohrmann commented on a change in pull request #13265: [FLINK-19055] Wait less time for all memory GC in tests (MemoryManager#verifyEmpty)
tillrohrmann commented on a change in pull request #13265: URL: https://github.com/apache/flink/pull/13265#discussion_r478863140 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/memory/UnsafeMemoryBudget.java ## @@ -36,16 +36,20 @@ */ class UnsafeMemoryBudget { private static final int MAX_SLEEPS = 11; // 2^11 - 1 = (2 x 1024) - 1 ms ~ 2 s total sleep duration - private static final int MAX_SLEEPS_VERIFY_EMPTY = 17; // 2^17 - 1 = (128 x 1024) - 1 ms ~ 2 min total sleep duration + static final int MAX_SLEEPS_VERIFY_EMPTY_FOR_TESTS = 10; // 2^10 - 1 = (1 x 1024) - 1 ms ~ 1 s total sleep duration Review comment: Testing code should not be contained in production code. ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/memory/MemoryManager.java ## @@ -89,12 +89,14 @@ * * @param memorySize The total size of the off-heap memory to be managed by this memory manager. * @param pageSize The size of the pages handed out by the memory manager. +* @param verifyEmptyWaitGcMaxSleeps defines how long to wait for GC of all allocated memory to check for memory leaks, +* see also {@link UnsafeMemoryBudget}. Review comment: Hmm, the description says that it is the time to wait for GC but isn't is the number of exponential retries? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Issue Comment Deleted] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nicholas Jiang updated FLINK-19064: --- Comment: was deleted (was: [~ZhuShang] Please pay attention to the defintion of method openInputFormat, which javadoc mentions that Resources should be allocated in this method. (e.g. database connections, cache, etc.).) > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.
flinkbot commented on pull request #13273: URL: https://github.com/apache/flink/pull/13273#issuecomment-682353427 ## CI report: * 0134fa06171742eb8ae840b2aced9530232bdb04 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers
flinkbot edited a comment on pull request #13272: URL: https://github.com/apache/flink/pull/13272#issuecomment-682346986 ## CI report: * 3ef9f728f6a8005a7f91846bf32df96ee718c626 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5952) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error
flinkbot edited a comment on pull request #13268: URL: https://github.com/apache/flink/pull/13268#issuecomment-682283788 ## CI report: * 5357812df45c42860e95630891367a410f6bc34b Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5944) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-19061) HiveCatalog fails to get partition column stats if partition value contains special characters
[ https://issues.apache.org/jira/browse/FLINK-19061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186297#comment-17186297 ] Jingsong Lee commented on FLINK-19061: -- master: f924a665b6f0bff815f87a221495b101b035315e > HiveCatalog fails to get partition column stats if partition value contains > special characters > -- > > Key: FLINK-19061 > URL: https://issues.apache.org/jira/browse/FLINK-19061 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Reporter: Rui Li >Assignee: Rui Li >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] JingsongLi merged pull request #13264: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…
JingsongLi merged pull request #13264: URL: https://github.com/apache/flink/pull/13264 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-19061) HiveCatalog fails to get partition column stats if partition value contains special characters
[ https://issues.apache.org/jira/browse/FLINK-19061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee reassigned FLINK-19061: Assignee: Rui Li > HiveCatalog fails to get partition column stats if partition value contains > special characters > -- > > Key: FLINK-19061 > URL: https://issues.apache.org/jira/browse/FLINK-19061 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Reporter: Rui Li >Assignee: Rui Li >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] JingsongLi commented on pull request #13264: [FLINK-19061][hive] HiveCatalog fails to get partition column stats i…
JingsongLi commented on pull request #13264: URL: https://github.com/apache/flink/pull/13264#issuecomment-682350350 Can you create a PR for release-1.11? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-19042) HiveTableSourceITCase fails if object reuse is enabled
[ https://issues.apache.org/jira/browse/FLINK-19042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee closed FLINK-19042. Assignee: Rui Li Resolution: Fixed master: 9e0964b81e502b266832fe2c271a1f8afdda5e5a > HiveTableSourceITCase fails if object reuse is enabled > -- > > Key: FLINK-19042 > URL: https://issues.apache.org/jira/browse/FLINK-19042 > Project: Flink > Issue Type: Test > Components: Connectors / Hive, Tests >Reporter: Rui Li >Assignee: Rui Li >Priority: Major > Labels: pull-request-available > > {{testNonPartitionStreamingSourceWithVectorizedReader}} fails because print > table sink cannot process {{ColumnarRowData}} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-19042) HiveTableSourceITCase fails if object reuse is enabled
[ https://issues.apache.org/jira/browse/FLINK-19042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee updated FLINK-19042: - Fix Version/s: 1.12.0 > HiveTableSourceITCase fails if object reuse is enabled > -- > > Key: FLINK-19042 > URL: https://issues.apache.org/jira/browse/FLINK-19042 > Project: Flink > Issue Type: Test > Components: Connectors / Hive, Tests >Reporter: Rui Li >Assignee: Rui Li >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0 > > > {{testNonPartitionStreamingSourceWithVectorizedReader}} fails because print > table sink cannot process {{ColumnarRowData}} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] JingsongLi merged pull request #13238: [FLINK-19042][hive] Remove print table sink from HiveTableSourceITCas…
JingsongLi merged pull request #13238: URL: https://github.com/apache/flink/pull/13238 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.
flinkbot commented on pull request #13273: URL: https://github.com/apache/flink/pull/13273#issuecomment-682347761 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 0134fa06171742eb8ae840b2aced9530232bdb04 (Fri Aug 28 06:16:50 UTC 2020) ✅no warnings Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-18801) Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section
[ https://issues.apache.org/jira/browse/FLINK-18801?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-18801: --- Labels: pull-request-available (was: ) > Add a "10 minutes to Table API" document under the "Python API" -> "User > Guide" -> "Table API" section > --- > > Key: FLINK-18801 > URL: https://issues.apache.org/jira/browse/FLINK-18801 > Project: Flink > Issue Type: Sub-task >Reporter: Wei Zhong >Assignee: Wei Zhong >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-19071) Some Hive window functions are not supported
[ https://issues.apache.org/jira/browse/FLINK-19071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rui Li updated FLINK-19071: --- Description: Including the following: {noformat} FIRST_VALUE and LAST_VALUE that take two parameters NTILE CUME_DIST PERCENT_RANK {noformat} was: Including the following: {noformat} FIRST_VALUE and LAST_VALUE that take two parameters NTILE CUME_DIST {noformat} > Some Hive window functions are not supported > > > Key: FLINK-19071 > URL: https://issues.apache.org/jira/browse/FLINK-19071 > Project: Flink > Issue Type: Task > Components: Connectors / Hive >Reporter: Rui Li >Priority: Major > > Including the following: > {noformat} > FIRST_VALUE and LAST_VALUE that take two parameters > NTILE > CUME_DIST > PERCENT_RANK > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers
flinkbot commented on pull request #13272: URL: https://github.com/apache/flink/pull/13272#issuecomment-682346986 ## CI report: * 3ef9f728f6a8005a7f91846bf32df96ee718c626 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] WeiZhong94 opened a new pull request #13273: [FLINK-18801][docs][python] Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.
WeiZhong94 opened a new pull request #13273: URL: https://github.com/apache/flink/pull/13273 ## What is the purpose of the change *This pull request adds a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.* ## Brief change log - *Add a "10 minutes to Table API" document under the "Python API" -> "User Guide" -> "Table API" section.* ## Verifying this change This change is a trivial rework / code cleanup without any test coverage. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: ( no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (docs) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
flinkbot edited a comment on pull request #13271: URL: https://github.com/apache/flink/pull/13271#issuecomment-682323547 ## CI report: * dcfea8ba4f522e9cd1a87022eadb10438902f22c Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5950) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13249: [FLINK-14087][datastream] Clone the StreamPartitioner to avoid being shared at runtime.
flinkbot edited a comment on pull request #13249: URL: https://github.com/apache/flink/pull/13249#issuecomment-680775587 ## CI report: * 07e11ae0c700a4172a854ae4fd108b483bde6003 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5947) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-19071) Some Hive window functions are not supported
[ https://issues.apache.org/jira/browse/FLINK-19071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rui Li updated FLINK-19071: --- Description: Including the following: {noformat} FIRST_VALUE and LAST_VALUE that take two parameters NTILE CUME_DIST {noformat} > Some Hive window functions are not supported > > > Key: FLINK-19071 > URL: https://issues.apache.org/jira/browse/FLINK-19071 > Project: Flink > Issue Type: Task > Components: Connectors / Hive >Reporter: Rui Li >Priority: Major > > Including the following: > {noformat} > FIRST_VALUE and LAST_VALUE that take two parameters > NTILE > CUME_DIST > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers
flinkbot commented on pull request #13272: URL: https://github.com/apache/flink/pull/13272#issuecomment-682341723 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 3ef9f728f6a8005a7f91846bf32df96ee718c626 (Fri Aug 28 05:58:42 UTC 2020) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] klion26 commented on a change in pull request #13225: [FLINK-18974][docs-zh]Translate the 'User-Defined Functions' page of "Application Development's DataStream API" into Chinese
klion26 commented on a change in pull request #13225: URL: https://github.com/apache/flink/pull/13225#discussion_r478844519 ## File path: docs/dev/user_defined_functions.zh.md ## @@ -147,95 +153,77 @@ data.map (new RichMapFunction[String, Int] { -Rich functions provide, in addition to the user-defined function (map, -reduce, etc), four methods: `open`, `close`, `getRuntimeContext`, and -`setRuntimeContext`. These are useful for parameterizing the function -(see [Passing Parameters to Functions]({{ site.baseurl }}/dev/batch/index.html#passing-parameters-to-functions)), -creating and finalizing local state, accessing broadcast variables (see -[Broadcast Variables]({{ site.baseurl }}/dev/batch/index.html#broadcast-variables)), and for accessing runtime -information such as accumulators and counters (see -[Accumulators and Counters](#accumulators--counters)), and information -on iterations (see [Iterations]({{ site.baseurl }}/dev/batch/iterations.html)). +除了用户自定义的功能(map,reduce 等),Rich functions 还提供了四个方法:`open`、`close`、`getRuntimeContext` 和 +`setRuntimeContext`。这些对于参数化功能很有用 +(参阅 [给函数传递参数]({{ site.baseurl }}/zh/dev/batch/index.html#passing-parameters-to-functions)), +创建和最终确定本地状态,访问广播变量(参阅 +[广播变量]({{ site.baseurl }}/zh/dev/batch/index.html#broadcast-variables)),以及访问运行时信息,例如累加器和计数器(参阅 +[累加器和计数器](#累加器和计数器)),以及迭代器的相关信息(参阅 [迭代器]({{ site.baseurl }}/zh/dev/batch/iterations.html))。 {% top %} -## Accumulators & Counters + -Accumulators are simple constructs with an **add operation** and a **final accumulated result**, -which is available after the job ended. +## 累加器和计数器 -The most straightforward accumulator is a **counter**: You can increment it using the -```Accumulator.add(V value)``` method. At the end of the job Flink will sum up (merge) all partial -results and send the result to the client. Accumulators are useful during debugging or if you -quickly want to find out more about your data. +累加器是具有**加法运算**和**最终累加结果**的一种简单结构,可在作业结束后使用。 -Flink currently has the following **built-in accumulators**. Each of them implements the -{% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/Accumulator.java "Accumulator" %} -interface. +最简单的累加器就是**计数器**: 你可以使用 +```Accumulator.add(V value)``` 方法将其递增。在作业结束时,Flink 会汇总(合并)所有部分的结果并将其发送给客户端。 +在调试过程中或在你想快速了解有关数据更多信息时,累加器作用很大。 + +Flink 目前有如下**内置累加器**。每个都实现了 +{% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/Accumulator.java "累加器" %} +接口。 - {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/IntCounter.java "__IntCounter__" %}, {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/LongCounter.java "__LongCounter__" %} - and {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/DoubleCounter.java "__DoubleCounter__" %}: - See below for an example using a counter. -- {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java "__Histogram__" %}: - A histogram implementation for a discrete number of bins. Internally it is just a map from Integer - to Integer. You can use this to compute distributions of values, e.g. the distribution of - words-per-line for a word count program. + 和 {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/DoubleCounter.java "__DoubleCounter__" %}: + 有关使用计数器的示例,请参见下文。 +- {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/accumulators/Histogram.java "__直方图__" %}: + 离散数量的柱状直方图实现。在内部,它只是整形到整形的映射。你可以使用它来计算值的分布,例如,单词计数程序的每行单词的分布情况。 -__How to use accumulators:__ +__如何使用累加器:__ -First you have to create an accumulator object (here a counter) in the user-defined transformation -function where you want to use it. +首先,你要在需要使用累加器的用户自定义的转换函数中创建一个累加器对象(此处是计数器)。 Review comment: ```suggestion 首先,在需要使用累加器的用户自定义函数中创建一个累加器对象(此处是计数器)。 ``` ## File path: docs/dev/user_defined_functions.zh.md ## @@ -147,95 +153,77 @@ data.map (new RichMapFunction[String, Int] { -Rich functions provide, in addition to the user-defined function (map, -reduce, etc), four methods: `open`, `close`, `getRuntimeContext`, and -`setRuntimeContext`. These are useful for parameterizing the function -(see [Passing Parameters to Functions]({{ site.baseurl }}/dev/batch/index.html#passing-parameters-to-functions)), -creating and finalizing local state, accessing broadcast variables (see -[Broadcast Variables]({{ site.baseurl }}/dev/batch/index.html#broadcast-variables)), and for accessing runtime -information such as accumulators and counters (see -[Accumulators and Counters](#accumulators--counters)), and information -on iterations (see [Iterations]({{ site.baseurl }}/dev/batch/iterations.html)). +除了用户自定义的功能(map,reduce 等),Rich functions 还提供了四个方法:`open`、`close`、`getRuntimeContext` 和 +`setRuntimeContext`。这些对于参数化功能很有用 +(参阅 [给函数传递参数]({{ site.baseurl }}/zh/dev/batch/index.html#passing-parameters-to-functi
[jira] [Updated] (FLINK-18695) Allow NettyBufferPool to allocate heap buffers
[ https://issues.apache.org/jira/browse/FLINK-18695?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-18695: --- Labels: pull-request-available (was: ) > Allow NettyBufferPool to allocate heap buffers > -- > > Key: FLINK-18695 > URL: https://issues.apache.org/jira/browse/FLINK-18695 > Project: Flink > Issue Type: Improvement > Components: Runtime / Network >Reporter: Chesnay Schepler >Assignee: Yun Gao >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0 > > > in 4.1.43 netty made a change to their SslHandler to always use heap buffers > for JDK SSLEngine implementations, to avoid an additional memory copy. > However, our {{NettyBufferPool}} forbids heap buffer allocations. > We will either have to allow heap buffer allocations, or create a custom > SslHandler implementation that does not use heap buffers (although this seems > ill-adviced?). > /cc [~sewen] [~uce] [~NicoK] [~zjwang] [~pnowojski] -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.
flinkbot edited a comment on pull request #13109: URL: https://github.com/apache/flink/pull/13109#issuecomment-671381707 ## CI report: * 3d854c62355f9049062a7ae6a908dcceecd9c213 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5932) * 87ec25bce258c9fc953084701dd3acf7d96ac9e2 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5951) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] gaoyunhaii opened a new pull request #13272: [FLINK-18695][network] Netty fakes heap buffer allocationn with direct buffers
gaoyunhaii opened a new pull request #13272: URL: https://github.com/apache/flink/pull/13272 ## What is the purpose of the change This PR modifies the `NettyBufferPool` to also allocate direct buffers for the heap buffer request. This enables us not to change the memory footprint when upgrade Netty to 4.1.50-FINAL. In the future we could further decide how to adjust the Netty memory management. ## Brief change log - 3ef9f728f6a8005a7f91846bf32df96ee718c626 returns direct buffers for all the heap buffer allocation methods. ## Verifying this change This change is a trivial rework / code cleanup without any test coverage. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): **no** - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: **no** - The serializers: **no** - The runtime per-record code paths (performance sensitive): **no** - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: **no** - The S3 file system connector: **no** ## Documentation - Does this pull request introduce a new feature? **no** - If yes, how is the feature documented? **not applicable** This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.
flinkbot edited a comment on pull request #13109: URL: https://github.com/apache/flink/pull/13109#issuecomment-671381707 ## CI report: * 3d854c62355f9049062a7ae6a908dcceecd9c213 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5932) * 87ec25bce258c9fc953084701dd3acf7d96ac9e2 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] liming30 commented on a change in pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.
liming30 commented on a change in pull request #13109: URL: https://github.com/apache/flink/pull/13109#discussion_r478836972 ## File path: flink-streaming-java/src/main/java/org/apache/flink/streaming/api/collector/selector/DirectedOutputsCollector.java ## @@ -0,0 +1,67 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.streaming.api.collector.selector; + +import org.apache.flink.streaming.api.operators.Output; +import org.apache.flink.streaming.runtime.streamrecord.StreamRecord; + +import java.util.Map; + +/** + * The selected outputs collector will send records to the default output, + * and output matching outputNames. + * + * @param The type of the elements that can be emitted. + */ +public class DirectedOutputsCollector implements SelectedOutputsCollector { + + private final Output>[] selectAllOutputs; + private final Map>[]> outputMap; + + public DirectedOutputsCollector( + Output>[] selectAllOutputs, + Map>[]> outputMap) { + this.selectAllOutputs = selectAllOutputs; + this.outputMap = outputMap; + } + + @Override + public boolean collect(Iterable outputNames, StreamRecord record) { + boolean emitted = false; + + if (selectAllOutputs.length > 0) { + collect(selectAllOutputs, record); + emitted = true; + } + + for (String outputName : outputNames) { + Output>[] outputList = outputMap.get(outputName); + if (outputList != null && outputList.length > 0) { + collect(outputList, record); + emitted = true; + } + } Review comment: In the old implementation via `set`, even if the same `output` appears multiple times in `outputNames`, it will only be sent once. Now it will send multiple times and I am not sure if this behavior is correct. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] liming30 commented on a change in pull request #13109: [FLINK-18808][runtime/metrics] Include side outputs in numRecordsOut metric.
liming30 commented on a change in pull request #13109: URL: https://github.com/apache/flink/pull/13109#discussion_r478431177 ## File path: flink-streaming-java/src/main/java/org/apache/flink/streaming/api/collector/selector/SelectedOutputsCollectorImpl.java ## @@ -0,0 +1,61 @@ +package org.apache.flink.streaming.api.collector.selector; + +import org.apache.flink.streaming.api.operators.Output; +import org.apache.flink.streaming.runtime.streamrecord.StreamRecord; + +import java.util.Map; + +/** + * The selected outputs collector will send records to the default output, + * and output matching outputNames. + * + * @param The type of the elements that can be emitted. + */ +public class SelectedOutputsCollectorImpl implements SelectedOutputsCollector { + + private final Output>[] selectAllOutputs; + private final Map>[]> outputMap; + + private final boolean objectReuse; + + public SelectedOutputsCollectorImpl( + Output>[] selectAllOutputs, + Map>[]> outputMap, + boolean objectReuse) { + this.selectAllOutputs = selectAllOutputs; + this.outputMap = outputMap; + this.objectReuse = objectReuse; + } + + @Override + public boolean collect(Iterable outputNames, StreamRecord record) { + boolean emitted = false; + + if (selectAllOutputs.length > 0) { + collect(selectAllOutputs, record); + emitted = true; + } + + for (String outputName : outputNames) { + Output>[] outputList = outputMap.get(outputName); + if (outputList != null && outputList.length > 0) { + collect(outputList, record); + emitted = true; + } + } Review comment: In the old implementation via `set`, even if the same `output` appears multiple times in `outputNames`, it will only be sent once. Now it will send multiple times and I am not sure if this behavior is correct. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] dianfu commented on pull request #13193: [FLINK-18918][python][docs] Add dedicated connector documentation for Python Table API
dianfu commented on pull request #13193: URL: https://github.com/apache/flink/pull/13193#issuecomment-682331400 @hequn8128 Thanks for the update. LGTM. @sjwiesman @morsapaes could you take a further look at of the latest PR? Thanks a lot! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
flinkbot edited a comment on pull request #13271: URL: https://github.com/apache/flink/pull/13271#issuecomment-682323547 ## CI report: * dcfea8ba4f522e9cd1a87022eadb10438902f22c Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5950) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
flinkbot commented on pull request #13271: URL: https://github.com/apache/flink/pull/13271#issuecomment-682323547 ## CI report: * dcfea8ba4f522e9cd1a87022eadb10438902f22c UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-19063) Support join late event from dimension table side in temporal table join
[ https://issues.apache.org/jira/browse/FLINK-19063?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186253#comment-17186253 ] Paul Lin commented on FLINK-19063: -- WRT event-time temporal table join, as an initial and naive thought, I think we can leverage the watermarks of both streams (may require some watermark mechanism changes). We should ensure the build side watermark is greater than the one of the probe size, if not, maybe we can keep the unjoined data of the probe side table before the build side watermark(plus allowed lateness) in the states. When a new element of the build side shows up, it triggers the unjoined data in the states to re-join, produce the join result (if any), and be removed from the states. The process would be similar to event-time interval join. > Support join late event from dimension table side in temporal table join > - > > Key: FLINK-19063 > URL: https://issues.apache.org/jira/browse/FLINK-19063 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Reporter: Leonard Xu >Priority: Major > > To join late event from dimension table side in temporal table join is a > common user case > from user-zh mail list[1][2]. > And another similar user case is how to enable the faster stream to wait the > slower stream in regular stream join[3]. > I think we can discuss how to support these user cases. > > > [1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html] > [2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585] > [3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436] -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] RocMarshal commented on a change in pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
RocMarshal commented on a change in pull request #13271: URL: https://github.com/apache/flink/pull/13271#discussion_r478825509 ## File path: docs/monitoring/logging.zh.md ## @@ -100,15 +106,13 @@ import org.slf4j.Logger Logger LOG = LoggerFactory.getLogger(Foobar.class) {% endhighlight %} -In order to benefit most from slf4j, it is recommended to use its placeholder mechanism. -Using placeholders allows to avoid unnecessary string constructions in case that the logging level is set so high that the message would not be logged. -The syntax of placeholders is the following: +为了最大限度地利用 slf4j,建议使用其占位符机制。使用占位符可以避免不必要的字符串构造,以防日志级别设置得太高而不会记录消息。占位符的语法如下: {% highlight java %} LOG.info("This message contains {} placeholders. {}", 2, "Yippie"); {% endhighlight %} -Placeholders can also be used in conjunction with exceptions which shall be logged. +占位符也可以与应记录的异常一起使用。 Review comment: ```suggestion 占位符也可以和要记录的异常一起使用。 ``` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
flinkbot commented on pull request #13271: URL: https://github.com/apache/flink/pull/13271#issuecomment-682319220 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit a50c1399395cbc32008f02ffdb0380c897f522a4 (Fri Aug 28 04:33:59 UTC 2020) ✅no warnings Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-19043) Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
[ https://issues.apache.org/jira/browse/FLINK-19043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-19043: --- Labels: Documentation Translation pull-request-available translation-zh (was: Documentation Translation translation-zh) > Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese > - > > Key: FLINK-19043 > URL: https://issues.apache.org/jira/browse/FLINK-19043 > Project: Flink > Issue Type: Improvement > Components: chinese-translation, Documentation >Affects Versions: 1.10.0, 1.10.1, 1.11.0, 1.10.2, 1.11.1 >Reporter: Roc Marshal >Assignee: Roc Marshal >Priority: Major > Labels: Documentation, Translation, pull-request-available, > translation-zh > > The page url is : > [Logging|https://ci.apache.org/projects/flink/flink-docs-release-1.11/monitoring/logging.html] > The markdown file location is : flink/docs/monitoring/logging.zh.md -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] RocMarshal opened a new pull request #13271: [FLINK-19043][docs-zh] Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese
RocMarshal opened a new pull request #13271: URL: https://github.com/apache/flink/pull/13271 ## What is the purpose of the change *Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese* ## Brief change log *Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese* - *The page url is : Logging https://ci.apache.org/projects/flink/flink-docs-release-1.11/monitoring/logging.html* - *The markdown file location is : flink/docs/monitoring/logging.zh.md* ## Verifying this change *Translate the 'Logging' page of 'Debugging & Monitoring' into Chinese* A pure translation work in the `Documentation` module. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: no - The serializers: no - The runtime per-record code paths (performance sensitive): no - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? no - If yes, how is the feature documented? not applicable & docs This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices
flinkbot edited a comment on pull request #13270: URL: https://github.com/apache/flink/pull/13270#issuecomment-682310308 ## CI report: * 04b974f56351d188b12bfe7c22a2fb8cb6e8f68e Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5948) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13249: [FLINK-14087][datastream] Clone the StreamPartitioner to avoid being shared at runtime.
flinkbot edited a comment on pull request #13249: URL: https://github.com/apache/flink/pull/13249#issuecomment-680775587 ## CI report: * 4eae231a2cd3edbb2ba34f7bef983f64a4c4fefa Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5915) * 07e11ae0c700a4172a854ae4fd108b483bde6003 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5947) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec
[ https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee closed FLINK-18900. Resolution: Fixed > HiveCatalog should error out when listing partitions with an invalid spec > - > > Key: FLINK-18900 > URL: https://issues.apache.org/jira/browse/FLINK-18900 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.11.1 >Reporter: Rui Li >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Take the following case as an example: > {code} > create table tbl (x int) partitioned by (p int); > alter table tbl add partition (p=1); > {code} > If we list partitions with partition spec {{foo=1}}, HiveCatalog returns > partition {{p=1}}, which is wrong. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec
[ https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186237#comment-17186237 ] Jingsong Lee commented on FLINK-18900: -- Revert interface modification for 1.11: a5767906916548ca51bf2d4b9e75c833ea6522a6 > HiveCatalog should error out when listing partitions with an invalid spec > - > > Key: FLINK-18900 > URL: https://issues.apache.org/jira/browse/FLINK-18900 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.11.1 >Reporter: Rui Li >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Take the following case as an example: > {code} > create table tbl (x int) partitioned by (p int); > alter table tbl add partition (p=1); > {code} > If we list partitions with partition spec {{foo=1}}, HiveCatalog returns > partition {{p=1}}, which is wrong. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] JingsongLi merged pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions
JingsongLi merged pull request #13269: URL: https://github.com/apache/flink/pull/13269 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices
flinkbot commented on pull request #13270: URL: https://github.com/apache/flink/pull/13270#issuecomment-682310308 ## CI report: * 04b974f56351d188b12bfe7c22a2fb8cb6e8f68e UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13249: [FLINK-14087][datastream] Clone the StreamPartitioner to avoid being shared at runtime.
flinkbot edited a comment on pull request #13249: URL: https://github.com/apache/flink/pull/13249#issuecomment-680775587 ## CI report: * 4eae231a2cd3edbb2ba34f7bef983f64a4c4fefa Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5915) * 07e11ae0c700a4172a854ae4fd108b483bde6003 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-19081) Deprecate TemporalTableFunction and Table#createTemporalTableFunction()
[ https://issues.apache.org/jira/browse/FLINK-19081?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Leonard Xu updated FLINK-19081: --- Summary: Deprecate TemporalTableFunction and Table#createTemporalTableFunction() (was: Deprecate TemporalTableFunction and Table$createTemporalTableFunction()) > Deprecate TemporalTableFunction and Table#createTemporalTableFunction() > --- > > Key: FLINK-19081 > URL: https://issues.apache.org/jira/browse/FLINK-19081 > Project: Flink > Issue Type: Sub-task >Reporter: Leonard Xu >Priority: Major > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19082) Add docs for temporal table and temporal table join
Leonard Xu created FLINK-19082: -- Summary: Add docs for temporal table and temporal table join Key: FLINK-19082 URL: https://issues.apache.org/jira/browse/FLINK-19082 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19081) Deprecate TemporalTableFunction and Table$createTemporalTableFunction()
Leonard Xu created FLINK-19081: -- Summary: Deprecate TemporalTableFunction and Table$createTemporalTableFunction() Key: FLINK-19081 URL: https://issues.apache.org/jira/browse/FLINK-19081 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19080) Materialize timeindicator data type in the right input of temporal join
Leonard Xu created FLINK-19080: -- Summary: Materialize timeindicator data type in the right input of temporal join Key: FLINK-19080 URL: https://issues.apache.org/jira/browse/FLINK-19080 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19079) Support row time deduplicate operator
Leonard Xu created FLINK-19079: -- Summary: Support row time deduplicate operator Key: FLINK-19079 URL: https://issues.apache.org/jira/browse/FLINK-19079 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19077) Improve process time temporal join operator
Leonard Xu created FLINK-19077: -- Summary: Improve process time temporal join operator Key: FLINK-19077 URL: https://issues.apache.org/jira/browse/FLINK-19077 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19078) Import rowtime join temporal operator
Leonard Xu created FLINK-19078: -- Summary: Import rowtime join temporal operator Key: FLINK-19078 URL: https://issues.apache.org/jira/browse/FLINK-19078 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19076) Import rule to deal Temporal Join condition
Leonard Xu created FLINK-19076: -- Summary: Import rule to deal Temporal Join condition Key: FLINK-19076 URL: https://issues.apache.org/jira/browse/FLINK-19076 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19075) Infer changelog trait for temporal join node
Leonard Xu created FLINK-19075: -- Summary: Infer changelog trait for temporal join node Key: FLINK-19075 URL: https://issues.apache.org/jira/browse/FLINK-19075 Project: Flink Issue Type: Bug Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19073) Improve streamExecTemporalJoinRule
Leonard Xu created FLINK-19073: -- Summary: Improve streamExecTemporalJoinRule Key: FLINK-19073 URL: https://issues.apache.org/jira/browse/FLINK-19073 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19074) Materialize timeindicator in the right input of temporal join
Leonard Xu created FLINK-19074: -- Summary: Materialize timeindicator in the right input of temporal join Key: FLINK-19074 URL: https://issues.apache.org/jira/browse/FLINK-19074 Project: Flink Issue Type: Bug Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices
flinkbot commented on pull request #13270: URL: https://github.com/apache/flink/pull/13270#issuecomment-682306916 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 04b974f56351d188b12bfe7c22a2fb8cb6e8f68e (Fri Aug 28 03:38:24 UTC 2020) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-19072) Import Temporal Table join rule
Leonard Xu created FLINK-19072: -- Summary: Import Temporal Table join rule Key: FLINK-19072 URL: https://issues.apache.org/jira/browse/FLINK-19072 Project: Flink Issue Type: Sub-task Reporter: Leonard Xu -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] sdlcwangsong opened a new pull request #13270: [hotfix] [javadocs] fix typo in TaskManagerServices
sdlcwangsong opened a new pull request #13270: URL: https://github.com/apache/flink/pull/13270 ## What is the purpose of the change *(For example: This pull request makes task deployment go through the blob server, rather than through RPC. That way we avoid re-transferring them on each deployment (during recovery).)* ## Brief change log *(for example:)* - *The TaskInfo is stored in the blob store on job creation time as a persistent artifact* - *Deployments RPC transmits only the blob storage reference* - *TaskManagers retrieve the TaskInfo from the blob cache* ## Verifying this change *(Please pick either of the following options)* This change is a trivial rework / code cleanup without any test coverage. *(or)* This change is already covered by existing tests, such as *(please describe tests)*. *(or)* This change added tests and can be verified as follows: *(example:)* - *Added integration tests for end-to-end deployment with large payloads (100MB)* - *Extended integration test for recovery after master (JobManager) failure* - *Added test that validates that TaskInfo is transferred only once across recoveries* - *Manually verified the change by running a 4 node cluser with 2 JobManagers and 4 TaskManagers, a stateful streaming program, and killing one JobManager and two TaskManagers during the execution, verifying that recovery happens correctly.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: no - The serializers: no - The runtime per-record code paths (performance sensitive): no - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? no - If yes, how is the feature documented? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.
flinkbot edited a comment on pull request #13263: URL: https://github.com/apache/flink/pull/13263#issuecomment-681892463 ## CI report: * 990aa0f33e2fa93d40ea011082fedc6eeb37be44 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5922) * 4f0f42441fadc0961dbdecc9f1cf5ced660387d4 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5946) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-19070) Hive connector should throw a meaningful exception if user reads/writes ACID tables
Rui Li created FLINK-19070: -- Summary: Hive connector should throw a meaningful exception if user reads/writes ACID tables Key: FLINK-19070 URL: https://issues.apache.org/jira/browse/FLINK-19070 Project: Flink Issue Type: Improvement Components: Connectors / Hive Reporter: Rui Li Fix For: 1.12.0 -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186232#comment-17186232 ] Nicholas Jiang edited comment on FLINK-19064 at 8/28/20, 3:22 AM: -- [~ZhuShang] Please pay attention to the defintion of method openInputFormat, which javadoc mentions that Resources should be allocated in this method. (e.g. database connections, cache, etc.). was (Author: nicholasjiang): [~ZhuShang] Please pay attention to the defintion of method openInputFormat, which javadoc mentions that Resources should be allocated in this method. (e.g. database connections, cache, etc.). > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186232#comment-17186232 ] Nicholas Jiang commented on FLINK-19064: [~ZhuShang] Please pay attention to the defintion of method openInputFormat, which javadoc mentions that Resources should be allocated in this method. (e.g. database connections, cache, etc.). > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19071) Some Hive window functions are not supported
Rui Li created FLINK-19071: -- Summary: Some Hive window functions are not supported Key: FLINK-19071 URL: https://issues.apache.org/jira/browse/FLINK-19071 Project: Flink Issue Type: Task Components: Connectors / Hive Reporter: Rui Li -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-19069) finalizeOnMaster takes too much time and client timeouts
[ https://issues.apache.org/jira/browse/FLINK-19069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186231#comment-17186231 ] Kenneth William Krugler commented on FLINK-19069: - I'd recently posted about a related issue to the dev mailing list, where I asked: {quote}[...] the default behavior of Hadoop’s FileOutputCommitter (with algorithm == 1) is to put files in task-specific sub-dirs. It’s depending on a post-completion “merge paths” action to be taken by what is (for Hadoop) the Application Master. I assume that when running on a real cluster, the HadoopOutputFormat.finalizeGlobal() method’s call to commitJob() would do this, but it doesn’t seem to be happening when I run locally. If I set the algorithm version to 2, then “merge paths” is handled by FileOutputCommitter immediately, and the HadoopOutputFormat code finds files in the expected location. Wondering if Flink should always be using version 2 of the algorithm, as that’s more performant when there are a lot of results (which is why it was added). {quote} > finalizeOnMaster takes too much time and client timeouts > > > Key: FLINK-19069 > URL: https://issues.apache.org/jira/browse/FLINK-19069 > Project: Flink > Issue Type: Improvement > Components: Runtime / Task >Affects Versions: 1.9.0 >Reporter: Jiayi Liao >Priority: Major > > Currently we execute {{finalizeOnMaster}} in JM's main thread, which may > stuck the JM for a very long time and client timeouts eventually. > For example, we'd like to write data to HDFS and commit files on JM, which > takes more than ten minutes to commit tens of thousands files. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.
flinkbot edited a comment on pull request #13263: URL: https://github.com/apache/flink/pull/13263#issuecomment-681892463 ## CI report: * 990aa0f33e2fa93d40ea011082fedc6eeb37be44 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5922) * 4f0f42441fadc0961dbdecc9f1cf5ced660387d4 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-16824) FLIP-132 Temporal Table DDL and Temporal Table Join
[ https://issues.apache.org/jira/browse/FLINK-16824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Leonard Xu updated FLINK-16824: --- Summary: FLIP-132 Temporal Table DDL and Temporal Table Join (was: Creating Temporal Table Function via DDL) > FLIP-132 Temporal Table DDL and Temporal Table Join > --- > > Key: FLINK-16824 > URL: https://issues.apache.org/jira/browse/FLINK-16824 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Reporter: Konstantin Knauf >Assignee: Leonard Xu >Priority: Major > Fix For: 1.12.0 > > > Currently, a Temporal Table Function can only be created via the Table API or > indirectly via the configuration file of the SQL Client. > It would be great, if this was also possible in pure SQL via a DDL statement. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-19063) Support join late event from dimension table side in temporal table join
[ https://issues.apache.org/jira/browse/FLINK-19063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Leonard Xu updated FLINK-19063: --- Description: To join late event from dimension table side in temporal table join is a common user case from user-zh mail list[1][2]. And another similar user case is how to enable the faster stream to wait the slower stream in regular stream join[3]. I think we can discuss how to support these user cases. [1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html] [2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585] [3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436] was: To join late event from dimension table side in temporal table join is a common user case from user-zh mail list[1][3]. And another similar user case is how to enable the faster stream to wait the slower stream in regular stream join[3]. I think we can discuss how to support these user cases. [1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html] [2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585] [3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436] > Support join late event from dimension table side in temporal table join > - > > Key: FLINK-19063 > URL: https://issues.apache.org/jira/browse/FLINK-19063 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Reporter: Leonard Xu >Priority: Major > > To join late event from dimension table side in temporal table join is a > common user case > from user-zh mail list[1][2]. > And another similar user case is how to enable the faster stream to wait the > slower stream in regular stream join[3]. > I think we can discuss how to support these user cases. > > > [1][http://apache-flink.147419.n8.nabble.com/Flink-join-td6563.html] > [2][http://apache-flink.147419.n8.nabble.com/flinksql-mysql-td3584.html#a3585] > [3][http://apache-flink.147419.n8.nabble.com/Flink-sql-td4435.html#a4436] -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-19005) used metaspace grow on every execution
[ https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186221#comment-17186221 ] ShenDa edited comment on FLINK-19005 at 8/28/20, 2:57 AM: -- [~chesnay] Thanks for your detailed instruction. But I still think there's maybe something wrong in Flink. I find that the JdbcInputFormat & JdbcOutputFormat is key reason cause the Metaspace OOM, because the java.sql.DriverManager doesn't release the reference of the Driver. The DriverManager is loaded by java.internal.ClassLoader but the driver is loaded by ChildFisrtClassLoader, which means the ChildFirstClassLoader can't be garbage collected according analyzation of dump file. The following code is used by me to reproduce the issue and I use org.postgresql.Driver as jdbc Driver. {code:java} public static void main(String[] args) throws Exception { EnvironmentSettings envSettings = EnvironmentSettings.newInstance() .useBlinkPlanner() !origin-jdbc-inputformat.png! .inBatchMode() .build(); TableEnvironment tEnv = TableEnvironment.create(envSettings); tEnv.executeSql( "CREATE TABLE " + INPUT_TABLE + "(" + "id BIGINT," + "timestamp6_col TIMESTAMP(6)," + "timestamp9_col TIMESTAMP(6)," + "time_col TIME," + "real_col FLOAT," + "decimal_col DECIMAL(10, 4)" + ") WITH (" + " 'connector.type'='jdbc'," + " 'connector.url'='" + DB_URL + "'," + " 'connector.table'='" + INPUT_TABLE + "'," + " 'connector.USERNAME'='" + USERNAME + "'," + " 'connector.PASSWORD'='" + PASSWORD + "'" + ")" ); TableResult tableResult = tEnv.executeSql("SELECT timestamp6_col, decimal_col FROM " + INPUT_TABLE); tableResult.collect(); } {code} And below diagram shows the Metaspace usage constantly growing up, and finally TaskManager will be offline. !origin-jdbc-inputformat.png! Additional, I try to fix this issue by appending the following code to the function closeInputFormat() which can finally trigger garbage collect in Metaspace. {code:java} try{ final Enumeration drivers = DriverManager.getDrivers(); while (drivers.hasMoreElements()) { DriverManager.deregisterDriver(drivers.nextElement()); } } catch (SQLException se) { LOG.info("Inputformat couldn't be closed - " + se.getMessage()); } {code} The following diagram shows the usage of Metaspace will be decreased. !modified-jdbc-inputformat.png! So, do you think it's a flink problem, and should we create a new issue to fix. was (Author: dadashen): [~chesnay] Thanks for your detailed instruction. But I still think there's maybe something wrong in Flink. I find that the JdbcInputFormat & JdbcOutputFormat is key reason cause the Metaspace OOM, because the java.sql.DriverManager doesn't release the reference of the Driver. The DriverManager is loaded by java.internal.ClassLoader but the driver is loaded by ChildFisrtClassLoader, which means the ChildFirstClassLoader can't be garbage collected according analyzation of dump file. The following code is used by me to reproduce the issue and I use org.postgresql.Driver as jdbc Driver. {code:java} public static void main(String[] args) throws Exception { EnvironmentSettings envSettings = EnvironmentSettings.newInstance() .useBlinkPlanner() !origin-jdbc-inputformat.png! .inBatchMode() .build(); TableEnvironment tEnv = TableEnvironment.create(envSettings); tEnv.executeSql( "CREATE TABLE " + INPUT_TABLE + "(" + "id BIGINT," + "timestamp6_col TIMESTAMP(6)," + "timestamp9_col TIMESTAMP(6)," + "time_col TIME," + "real_col FLOAT," + "decimal_col DECIMAL(10, 4)" + ") WITH (" + " 'connector.type'='jdbc'," + " 'connector.url'='" + DB_URL + "'," + " 'connector.table'='" + INPUT_TABLE + "'," +
[GitHub] [flink] xintongsong commented on pull request #13263: [FLINK-17273][runtime] ActiveResourceManager closes task manager connection on worker terminated.
xintongsong commented on pull request #13263: URL: https://github.com/apache/flink/pull/13263#issuecomment-682296119 Thanks for the review, @tillrohrmann This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186224#comment-17186224 ] Nicholas Jiang commented on FLINK-19064: [~jark] I guess the meaning of [~ZhuShang] is that the javadoc of the method configure() is Creates a \{@link Scan} object and opens the \{@link HTable} connection, which means that HBase connection is defined to open in configure(). > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186224#comment-17186224 ] Nicholas Jiang edited comment on FLINK-19064 at 8/28/20, 2:51 AM: -- [~jark] I guess the meaning of [~ZhuShang] is that the javadoc of the method configure() is Creates a \{@link Scan} object and opens the \{@link HTable} connection, which means that HBase connection is defined to open in configure(). was (Author: nicholasjiang): [~jark] I guess the meaning of [~ZhuShang] is that the javadoc of the method configure() is Creates a \{@link Scan} object and opens the \{@link HTable} connection, which means that HBase connection is defined to open in configure(). > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-19005) used metaspace grow on every execution
[ https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186221#comment-17186221 ] ShenDa commented on FLINK-19005: [~chesnay] Thanks for your detailed instruction. But I still think there's maybe something wrong in Flink. I find that the JdbcInputFormat & JdbcOutputFormat is key reason cause the Metaspace OOM, because the java.sql.DriverManager doesn't release the reference of the Driver. The DriverManager is loaded by java.internal.ClassLoader but the driver is loaded by ChildFisrtClassLoader, which means the ChildFirstClassLoader can't be garbage collected according analyzation of dump file. The following code is used by me to reproduce the issue and I use org.postgresql.Driver as jdbc Driver. {code:java} public static void main(String[] args) throws Exception { EnvironmentSettings envSettings = EnvironmentSettings.newInstance() .useBlinkPlanner() !origin-jdbc-inputformat.png! .inBatchMode() .build(); TableEnvironment tEnv = TableEnvironment.create(envSettings); tEnv.executeSql( "CREATE TABLE " + INPUT_TABLE + "(" + "id BIGINT," + "timestamp6_col TIMESTAMP(6)," + "timestamp9_col TIMESTAMP(6)," + "time_col TIME," + "real_col FLOAT," + "decimal_col DECIMAL(10, 4)" + ") WITH (" + " 'connector.type'='jdbc'," + " 'connector.url'='" + DB_URL + "'," + " 'connector.table'='" + INPUT_TABLE + "'," + " 'connector.USERNAME'='" + USERNAME + "'," + " 'connector.PASSWORD'='" + PASSWORD + "'" + ")" ); TableResult tableResult = tEnv.executeSql("SELECT timestamp6_col, decimal_col FROM " + INPUT_TABLE); tableResult.collect(); } {code} And below diagram shows the Metaspace usage constantly growing up, and finally TaskManager will be offline. !origin-jdbc-inputformat.png! Additional, I try to fix this issue by appending the following code to the function closeInputFormat() which can finally trigger garbage collect in Metaspace. {code:java} try{ final Enumeration drivers = DriverManager.getDrivers(); while (drivers.hasMoreElements()) { DriverManager.deregisterDriver(drivers.nextElement()); } } catch (SQLException se) { LOG.info("Inputformat couldn't be closed - " + se.getMessage()); } {code} The following diagram shows the usage of Metaspace will be decreased. !modified-jdbc-inputformat.png! > used metaspace grow on every execution > -- > > Key: FLINK-19005 > URL: https://issues.apache.org/jira/browse/FLINK-19005 > Project: Flink > Issue Type: Bug > Components: Client / Job Submission, Runtime / Configuration, > Runtime / Coordination >Affects Versions: 1.11.1 >Reporter: Guillermo Sánchez >Assignee: Chesnay Schepler >Priority: Major > Attachments: heap_dump_after_10_executions.zip, > heap_dump_after_1_execution.zip, heap_dump_echo_lee.tar.xz, > modified-jdbc-inputformat.png, origin-jdbc-inputformat.png > > > Hi ! > Im running a 1.11.1 flink cluster, where I execute batch jobs made with > DataSet API. > I submit these jobs every day to calculate daily data. > In every execution, cluster's used metaspace increase by 7MB and its never > released. > This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i > need to restart the cluster to clean the metaspace > taskmanager.memory.jvm-metaspace.size is set to 512mb > Any idea of what could be causing this metaspace grow and why is it not > released ? > > > === Summary == > > Case 1, reported by [~gestevez]: > * Flink 1.11.1 > * Java 11 > * Maximum Metaspace size set to 512mb > * Custom Batch job, submitted daily > * Requires restart every 15 days after an OOM > Case 2, reported by [~Echo Lee]: > * Flink 1.11.0 > * Java 11 > * G1GC > * WordCount Batch job, submitted every second / every 5 minutes > * eventually fails TaskExecutor with OOM > Case 3, reported by [~DaDaShen] > * Flink 1.11.0 > * Java 11 > * WordCount Batch job, submitted every 5 seconds > * growing Metaspace, eventually OOM > -- This message was se
[jira] [Updated] (FLINK-19005) used metaspace grow on every execution
[ https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ShenDa updated FLINK-19005: --- Attachment: modified-jdbc-inputformat.png > used metaspace grow on every execution > -- > > Key: FLINK-19005 > URL: https://issues.apache.org/jira/browse/FLINK-19005 > Project: Flink > Issue Type: Bug > Components: Client / Job Submission, Runtime / Configuration, > Runtime / Coordination >Affects Versions: 1.11.1 >Reporter: Guillermo Sánchez >Assignee: Chesnay Schepler >Priority: Major > Attachments: heap_dump_after_10_executions.zip, > heap_dump_after_1_execution.zip, heap_dump_echo_lee.tar.xz, > modified-jdbc-inputformat.png, origin-jdbc-inputformat.png > > > Hi ! > Im running a 1.11.1 flink cluster, where I execute batch jobs made with > DataSet API. > I submit these jobs every day to calculate daily data. > In every execution, cluster's used metaspace increase by 7MB and its never > released. > This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i > need to restart the cluster to clean the metaspace > taskmanager.memory.jvm-metaspace.size is set to 512mb > Any idea of what could be causing this metaspace grow and why is it not > released ? > > > === Summary == > > Case 1, reported by [~gestevez]: > * Flink 1.11.1 > * Java 11 > * Maximum Metaspace size set to 512mb > * Custom Batch job, submitted daily > * Requires restart every 15 days after an OOM > Case 2, reported by [~Echo Lee]: > * Flink 1.11.0 > * Java 11 > * G1GC > * WordCount Batch job, submitted every second / every 5 minutes > * eventually fails TaskExecutor with OOM > Case 3, reported by [~DaDaShen] > * Flink 1.11.0 > * Java 11 > * WordCount Batch job, submitted every 5 seconds > * growing Metaspace, eventually OOM > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19069) finalizeOnMaster takes too much time and client timeouts
Jiayi Liao created FLINK-19069: -- Summary: finalizeOnMaster takes too much time and client timeouts Key: FLINK-19069 URL: https://issues.apache.org/jira/browse/FLINK-19069 Project: Flink Issue Type: Improvement Components: Runtime / Task Affects Versions: 1.9.0 Reporter: Jiayi Liao Currently we execute {{finalizeOnMaster}} in JM's main thread, which may stuck the JM for a very long time and client timeouts eventually. For example, we'd like to write data to HDFS and commit files on JM, which takes more than ten minutes to commit tens of thousands files. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions
flinkbot edited a comment on pull request #13269: URL: https://github.com/apache/flink/pull/13269#issuecomment-682288332 ## CI report: * c967e9837780e4ac82c668845bc35d1a38c47ee8 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5945) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186215#comment-17186215 ] Jark Wu commented on FLINK-19064: - Hi [~ZhuShang], why the connection should be created in {{configure()}}? > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Assigned] (FLINK-19064) HBaseRowDataInputFormat is leaking resources
[ https://issues.apache.org/jira/browse/FLINK-19064?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu reassigned FLINK-19064: --- Assignee: Nicholas Jiang > HBaseRowDataInputFormat is leaking resources > > > Key: FLINK-19064 > URL: https://issues.apache.org/jira/browse/FLINK-19064 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Nicholas Jiang >Priority: Critical > > {{HBaseRowDataInputFormat.configure()}} is calling {{connectToTable()}}, > which creates a connection to HBase that is not closed again. > A user reported this problem on the user@ list: > https://lists.apache.org/thread.html/ra04f6996eb50ee83aabd2ad0d50bec9afb6a924bfbb48ada3269c6d8%40%3Cuser.flink.apache.org%3E -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-19005) used metaspace grow on every execution
[ https://issues.apache.org/jira/browse/FLINK-19005?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ShenDa updated FLINK-19005: --- Attachment: origin-jdbc-inputformat.png > used metaspace grow on every execution > -- > > Key: FLINK-19005 > URL: https://issues.apache.org/jira/browse/FLINK-19005 > Project: Flink > Issue Type: Bug > Components: Client / Job Submission, Runtime / Configuration, > Runtime / Coordination >Affects Versions: 1.11.1 >Reporter: Guillermo Sánchez >Assignee: Chesnay Schepler >Priority: Major > Attachments: heap_dump_after_10_executions.zip, > heap_dump_after_1_execution.zip, heap_dump_echo_lee.tar.xz, > origin-jdbc-inputformat.png > > > Hi ! > Im running a 1.11.1 flink cluster, where I execute batch jobs made with > DataSet API. > I submit these jobs every day to calculate daily data. > In every execution, cluster's used metaspace increase by 7MB and its never > released. > This ends up with an OutOfMemoryError caused by Metaspace every 15 days and i > need to restart the cluster to clean the metaspace > taskmanager.memory.jvm-metaspace.size is set to 512mb > Any idea of what could be causing this metaspace grow and why is it not > released ? > > > === Summary == > > Case 1, reported by [~gestevez]: > * Flink 1.11.1 > * Java 11 > * Maximum Metaspace size set to 512mb > * Custom Batch job, submitted daily > * Requires restart every 15 days after an OOM > Case 2, reported by [~Echo Lee]: > * Flink 1.11.0 > * Java 11 > * G1GC > * WordCount Batch job, submitted every second / every 5 minutes > * eventually fails TaskExecutor with OOM > Case 3, reported by [~DaDaShen] > * Flink 1.11.0 > * Java 11 > * WordCount Batch job, submitted every 5 seconds > * growing Metaspace, eventually OOM > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-19068) Filter verbose pod events for KubernetesResourceManagerDriver
Xintong Song created FLINK-19068: Summary: Filter verbose pod events for KubernetesResourceManagerDriver Key: FLINK-19068 URL: https://issues.apache.org/jira/browse/FLINK-19068 Project: Flink Issue Type: Improvement Components: Deployment / Kubernetes Reporter: Xintong Song A status of a Kubernetes pod consists of many detailed fields. Currently, Flink receives pod {{MODIFIED}} events from the {{KubernetesPodsWatcher}} on every single change to these fields, many of which Flink does not care. The verbose events will not affect the functionality of Flink, but will pollute the logs with repeated messages, because Flink only looks into the fields it interested in and those fields are identical. E.g., when a task manager is stopped due to idle timeout, Flink receives 3 events: * MODIFIED: container terminated * MODIFIED: {{deletionGracePeriodSeconds}} changes from 30 to 0, which is a Kubernetes internal status change after containers are gracefully terminated * DELETED: Flink removes metadata of the terminated pod Among the 3 messages, Flink is only interested in the 1st MODIFIED message, but will try to process all of them because the container status is terminated. I propose to Filter the verbose events in {{KubernetesResourceManagerDriver.PodCallbackHandlerImpl}}, to only process the status changes interested by Flink. This probably requires recording the status of all living pods, to compare with the incoming events for detecting status changes. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions
flinkbot commented on pull request #13269: URL: https://github.com/apache/flink/pull/13269#issuecomment-682288332 ## CI report: * c967e9837780e4ac82c668845bc35d1a38c47ee8 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error
flinkbot edited a comment on pull request #13268: URL: https://github.com/apache/flink/pull/13268#issuecomment-682283788 ## CI report: * 5357812df45c42860e95630891367a410f6bc34b Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=5944) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-15719) Exceptions when using scala types directly with the State Process API
[ https://issues.apache.org/jira/browse/FLINK-15719?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186212#comment-17186212 ] Ying Z commented on FLINK-15719: I pull a request here, [https://github.com/apache/flink/pull/13266] could someone do a review, thanks. > Exceptions when using scala types directly with the State Process API > - > > Key: FLINK-15719 > URL: https://issues.apache.org/jira/browse/FLINK-15719 > Project: Flink > Issue Type: Bug > Components: API / State Processor >Affects Versions: 1.9.1 >Reporter: Ying Z >Assignee: Tzu-Li (Gordon) Tai >Priority: Major > Labels: pull-request-available > > I followed these steps to generate and read states: > # implements the example[1] `CountWindowAverage` in Scala(exactly same), and > run jobA => that makes good. > # execute `flink cancel -s ${JobID}` => savepoints was generated as expected. > # implements the example[2] `StatefulFunctionWithTime` in Scala(code below), > and run jobB => failed, exceptions shows that "Caused by: > org.apache.flink.util.StateMigrationException: The new key serializer must be > compatible." > ReaderFunction code as below: > {code:java} > // code placeholder > class ReaderFunction extends KeyedStateReaderFunction[Long, (Long, Long)] { > var countState: ValueState[(Long, Long)] = _ > override def open(parameters: Configuration): Unit = { > val stateDescriptor = new ValueStateDescriptor("average", > createTypeInformation[(Long, Long)]) > countState = getRuntimeContext().getState(stateDescriptor) > }override def readKey(key: Long, ctx: > KeyedStateReaderFunction.Context, out: Collector[(Long, Long)]): Unit = { > out.collect(countState.value()) > } > } > {code} > 1: > [https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/stream/state/state.html#using-managed-keyed-state] > > 2: > [https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/libs/state_processor_api.html#keyed-state] > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Closed] (FLINK-19060) Checkpoint not triggered when use broadcast stream
[ https://issues.apache.org/jira/browse/FLINK-19060?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] henvealf closed FLINK-19060. Resolution: Not A Bug > Checkpoint not triggered when use broadcast stream > -- > > Key: FLINK-19060 > URL: https://issues.apache.org/jira/browse/FLINK-19060 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.11.1 >Reporter: henvealf >Priority: Major > Attachments: image-2020-08-27-16-41-23-699.png, > image-2020-08-27-16-44-37-442.png, image-2020-08-27-16-45-28-134.png, > image-2020-08-27-16-51-10-512.png > > Original Estimate: 1h > Remaining Estimate: 1h > > Code: > !image-2020-08-27-16-51-10-512.png! > KafkaSourceConfig: > consumer.setStartFromGroupOffsets() > Web UI: > !image-2020-08-27-16-45-28-134.png! > Checkpoint always doesn't happen. Did I write something wrong? > Thanks! -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions
flinkbot commented on pull request #13269: URL: https://github.com/apache/flink/pull/13269#issuecomment-682286516 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit c967e9837780e4ac82c668845bc35d1a38c47ee8 (Fri Aug 28 02:17:35 UTC 2020) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] JingsongLi opened a new pull request #13269: [FLINK-18900][table] Revert the modification of Catalog.listPartitions
JingsongLi opened a new pull request #13269: URL: https://github.com/apache/flink/pull/13269 Revert the modification of Catalog.listPartitions for version Compatibility. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec
[ https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186205#comment-17186205 ] Rui Li commented on FLINK-18900: Perhaps we can revert the change in 1.11, given that 1.11 doesn't support SHOW PARTITIONS. > HiveCatalog should error out when listing partitions with an invalid spec > - > > Key: FLINK-18900 > URL: https://issues.apache.org/jira/browse/FLINK-18900 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.11.1 >Reporter: Rui Li >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Take the following case as an example: > {code} > create table tbl (x int) partitioned by (p int); > alter table tbl add partition (p=1); > {code} > If we list partitions with partition spec {{foo=1}}, HiveCatalog returns > partition {{p=1}}, which is wrong. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec
[ https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186204#comment-17186204 ] Jingsong Lee edited comment on FLINK-18900 at 8/28/20, 2:08 AM: We should not modify method throws exception for {{Catalog.listPartitions}} in release-1.11. was (Author: lzljs3620320): We should not modify method throws exception for {{Catalog.listPartitions}}. > HiveCatalog should error out when listing partitions with an invalid spec > - > > Key: FLINK-18900 > URL: https://issues.apache.org/jira/browse/FLINK-18900 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.11.1 >Reporter: Rui Li >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Take the following case as an example: > {code} > create table tbl (x int) partitioned by (p int); > alter table tbl add partition (p=1); > {code} > If we list partitions with partition spec {{foo=1}}, HiveCatalog returns > partition {{p=1}}, which is wrong. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec
[ https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17186204#comment-17186204 ] Jingsong Lee commented on FLINK-18900: -- We should not modify method throws exception for {{Catalog.listPartitions}}. > HiveCatalog should error out when listing partitions with an invalid spec > - > > Key: FLINK-18900 > URL: https://issues.apache.org/jira/browse/FLINK-18900 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.11.1 >Reporter: Rui Li >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Take the following case as an example: > {code} > create table tbl (x int) partitioned by (p int); > alter table tbl add partition (p=1); > {code} > If we list partitions with partition spec {{foo=1}}, HiveCatalog returns > partition {{p=1}}, which is wrong. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Reopened] (FLINK-18900) HiveCatalog should error out when listing partitions with an invalid spec
[ https://issues.apache.org/jira/browse/FLINK-18900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee reopened FLINK-18900: -- > HiveCatalog should error out when listing partitions with an invalid spec > - > > Key: FLINK-18900 > URL: https://issues.apache.org/jira/browse/FLINK-18900 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.11.1 >Reporter: Rui Li >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.12.0, 1.11.2 > > > Take the following case as an example: > {code} > create table tbl (x int) partitioned by (p int); > alter table tbl add partition (p=1); > {code} > If we list partitions with partition spec {{foo=1}}, HiveCatalog returns > partition {{p=1}}, which is wrong. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error
flinkbot commented on pull request #13268: URL: https://github.com/apache/flink/pull/13268#issuecomment-682283788 ## CI report: * 5357812df45c42860e95630891367a410f6bc34b UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-19067) FileNotFoundException when run flink examples on standby JobManager
JieFang.He created FLINK-19067: -- Summary: FileNotFoundException when run flink examples on standby JobManager Key: FLINK-19067 URL: https://issues.apache.org/jira/browse/FLINK-19067 Project: Flink Issue Type: Bug Affects Versions: 1.11.1 Reporter: JieFang.He 1、When run examples/batch/WordCount.jar on standby JobManager,it will fail with the exception: Caused by: java.io.FileNotFoundException: /data2/storageDir/default/blob/job_d29414828f614d5466e239be4d3889ac/blob_p-a2ebe1c5aa160595f214b4bd0f39d80e42ee2e93-f458f1c12dc023e78d25f191de1d7c4b (No such file or directory) at java.io.FileInputStream.open0(Native Method) at java.io.FileInputStream.open(FileInputStream.java:195) at java.io.FileInputStream.(FileInputStream.java:138) at org.apache.flink.core.fs.local.LocalDataInputStream.(LocalDataInputStream.java:50) at org.apache.flink.core.fs.local.LocalFileSystem.open(LocalFileSystem.java:143) at org.apache.flink.runtime.blob.FileSystemBlobStore.get(FileSystemBlobStore.java:105) at org.apache.flink.runtime.blob.FileSystemBlobStore.get(FileSystemBlobStore.java:87) at org.apache.flink.runtime.blob.BlobServer.getFileInternal(BlobServer.java:501) at org.apache.flink.runtime.blob.BlobServerConnection.get(BlobServerConnection.java:231) at org.apache.flink.runtime.blob.BlobServerConnection.run(BlobServerConnection.java:117) 2、Run examples success on other nodes 3、After run success on the other node, it can run success on the Standby JobManager. But run again will fail -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #13268: [FLINK-18988][hotfix][doc] Fix Flink Kafka Connector Dependency Error
flinkbot commented on pull request #13268: URL: https://github.com/apache/flink/pull/13268#issuecomment-682280167 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 5357812df45c42860e95630891367a410f6bc34b (Fri Aug 28 01:53:11 UTC 2020) **Warnings:** * Documentation files were touched, but no `.zh.md` files: Update Chinese documentation or file Jira ticket. Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org