[jira] [Commented] (FLINK-21030) Broken job restart for job with disjoint graph
[ https://issues.apache.org/jira/browse/FLINK-21030?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268409#comment-17268409 ] Zhu Zhu commented on FLINK-21030: - Agreed to trigger a global failover to bring FINISHED tasks back to RUNNING if stop-with-savepoint fails. Maybe right after that the stopped checkpoint scheduler is restarted in {{SchedulerBase#stopWithSavepoint()}}. > Broken job restart for job with disjoint graph > -- > > Key: FLINK-21030 > URL: https://issues.apache.org/jira/browse/FLINK-21030 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.11.2 >Reporter: Theo Diefenthal >Priority: Blocker > Fix For: 1.13.0, 1.11.4, 1.12.2 > > > Building on top of bugs: > https://issues.apache.org/jira/browse/FLINK-21028 > and https://issues.apache.org/jira/browse/FLINK-21029 : > I tried to stop a Flink application on YARN via savepoint which didn't > succeed due to a possible bug/racecondition in shutdown (Bug 21028). Due to > some reason, Flink attempted to restart the pipeline after the failure in > shutdown (21029). The bug here: > As I mentioned: My jobgraph is disjoint and the pipelines are fully isolated. > Lets say the original error occured in a single task of pipeline1. Flink then > restarted the entire pipeline1, but pipeline2 was shutdown successfully and > switched the state to FINISHED. > My job thus was in kind of an invalid state after the attempt to stopping: > One of two pipelines was running, the other was FINISHED. I guess this is > kind of a bug in the restarting behavior that only all connected components > of a graph are restarted, but the others aren't... -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-21030) Broken job restart for job with disjoint graph
[ https://issues.apache.org/jira/browse/FLINK-21030?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268409#comment-17268409 ] Zhu Zhu edited comment on FLINK-21030 at 1/20/21, 7:21 AM: --- Agreed to trigger a global failover to bring FINISHED tasks back to RUNNING if stop-with-savepoint fails. Maybe right after the stopped checkpoint scheduler is restarted in {{SchedulerBase#stopWithSavepoint()}}. was (Author: zhuzh): Agreed to trigger a global failover to bring FINISHED tasks back to RUNNING if stop-with-savepoint fails. Maybe right after that the stopped checkpoint scheduler is restarted in {{SchedulerBase#stopWithSavepoint()}}. > Broken job restart for job with disjoint graph > -- > > Key: FLINK-21030 > URL: https://issues.apache.org/jira/browse/FLINK-21030 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.11.2 >Reporter: Theo Diefenthal >Priority: Blocker > Fix For: 1.13.0, 1.11.4, 1.12.2 > > > Building on top of bugs: > https://issues.apache.org/jira/browse/FLINK-21028 > and https://issues.apache.org/jira/browse/FLINK-21029 : > I tried to stop a Flink application on YARN via savepoint which didn't > succeed due to a possible bug/racecondition in shutdown (Bug 21028). Due to > some reason, Flink attempted to restart the pipeline after the failure in > shutdown (21029). The bug here: > As I mentioned: My jobgraph is disjoint and the pipelines are fully isolated. > Lets say the original error occured in a single task of pipeline1. Flink then > restarted the entire pipeline1, but pipeline2 was shutdown successfully and > switched the state to FINISHED. > My job thus was in kind of an invalid state after the attempt to stopping: > One of two pipelines was running, the other was FINISHED. I guess this is > kind of a bug in the restarting behavior that only all connected components > of a graph are restarted, but the others aren't... -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] yuruguo removed a comment on pull request #14671: [hotfix][docs] Fix typo in 'docs/dev/table/functions/udfs.md & udfs.zh.md'.
yuruguo removed a comment on pull request #14671: URL: https://github.com/apache/flink/pull/14671#issuecomment-763391201 LGTM. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14704: [FLINK-21044][connectors/jdbc] Use more random bytes in Xid
flinkbot edited a comment on pull request #14704: URL: https://github.com/apache/flink/pull/14704#issuecomment-763375645 ## CI report: * d37ad7dc90d2f07f30b9db9fc5185d591f1189f8 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12267) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
flinkbot edited a comment on pull request #14699: URL: https://github.com/apache/flink/pull/14699#issuecomment-763287840 ## CI report: * 23ac68d1604e6f5e38af38caa73d11a8dc76ceb9 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12268) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] yuruguo commented on pull request #14671: [hotfix][docs] Fix typo in 'docs/dev/table/functions/udfs.md & udfs.zh.md'.
yuruguo commented on pull request #14671: URL: https://github.com/apache/flink/pull/14671#issuecomment-763391201 LGTM. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14704: [FLINK-21044][connectors/jdbc] Use more random bytes in Xid
flinkbot edited a comment on pull request #14704: URL: https://github.com/apache/flink/pull/14704#issuecomment-763375645 ## CI report: * d37ad7dc90d2f07f30b9db9fc5185d591f1189f8 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12267) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
flinkbot edited a comment on pull request #14699: URL: https://github.com/apache/flink/pull/14699#issuecomment-763287840 ## CI report: * ddd72a9067d9c7b0fed65b26947d00f530d53310 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12258) * 23ac68d1604e6f5e38af38caa73d11a8dc76ceb9 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] RocMarshal edited a comment on pull request #14671: [hotfix][docs] Fix typo in 'docs/dev/table/functions/udfs.md & udfs.zh.md'.
RocMarshal edited a comment on pull request #14671: URL: https://github.com/apache/flink/pull/14671#issuecomment-762580819 @appleyuchi Hi, Could you help me to review this PR if you have any free time? Thanks! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] RocMarshal commented on pull request #14702: [FLINK-21042][docs] Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
RocMarshal commented on pull request #14702: URL: https://github.com/apache/flink/pull/14702#issuecomment-763384117 @appleyuchi Hi, Could you help me to review this PR if you have any free time? Thanks! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] leonardBang commented on pull request #14684: [FLINK-20460][Connector-HBase] Support async lookup for HBase connector
leonardBang commented on pull request #14684: URL: https://github.com/apache/flink/pull/14684#issuecomment-763383818 > @leonardBang My current approach to conflict is not friendly and resulted in the inclusion of other commits. So I want to create a new branch and apply for a new pull request. Do you think it is OK? You don't need to raise up a new PR, you can prepare your new branch(e.g: `branch_a`) which based on latest master and appended with a single commit that contains your change, and then (1) delete the `anlen321:FLINK-20460` in your local `git branch -D FLINK-20460`, (3) checkout a new branch named `FLINK-20460` from the `branch_a`,(4) push the `FLINK-20460` branch to remote `git push origin FLINK-20460`, the PR will automatically update This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-21048) Refactor documentation related to switch memory allocator
Yun Tang created FLINK-21048: Summary: Refactor documentation related to switch memory allocator Key: FLINK-21048 URL: https://issues.apache.org/jira/browse/FLINK-21048 Project: Flink Issue Type: Improvement Components: Documentation Affects Versions: 1.13.0, 1.12.2 Reporter: Yun Tang Assignee: Yun Tang Since we decide to change the switch of memory allocator from command to environment variable in FLINK-21034, we should also change related documentation. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21047) Expose the correct registered/free resources information in SlotManager
[ https://issues.apache.org/jira/browse/FLINK-21047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yangze Guo updated FLINK-21047: --- Affects Version/s: 1.11.3 1.12.1 > Expose the correct registered/free resources information in SlotManager > --- > > Key: FLINK-21047 > URL: https://issues.apache.org/jira/browse/FLINK-21047 > Project: Flink > Issue Type: Improvement >Affects Versions: 1.11.3, 1.12.1 >Reporter: Yangze Guo >Priority: Major > > In FLINK-16640, we extend ResourceOverview and TaskManager(Details)Info for > registered/free resources. However, the implementation is based on the > assumption that all the registered task executors are homogeneous with the > default resource spec. This assumption will be broken in standalone mode when > user manually starts heterogeneous task executors. > We need to calculate the registered/free resources according to the exact > {{defaultSlotResourceProfile}} and {{totalResourceProfile}} of > \{{TaskExecutorRegistration}}s. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21047) Expose the correct registered/free resources information in SlotManager
[ https://issues.apache.org/jira/browse/FLINK-21047?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yangze Guo updated FLINK-21047: --- Component/s: Runtime / Metrics Runtime / Coordination > Expose the correct registered/free resources information in SlotManager > --- > > Key: FLINK-21047 > URL: https://issues.apache.org/jira/browse/FLINK-21047 > Project: Flink > Issue Type: Improvement > Components: Runtime / Coordination, Runtime / Metrics >Affects Versions: 1.11.3, 1.12.1 >Reporter: Yangze Guo >Priority: Major > > In FLINK-16640, we extend ResourceOverview and TaskManager(Details)Info for > registered/free resources. However, the implementation is based on the > assumption that all the registered task executors are homogeneous with the > default resource spec. This assumption will be broken in standalone mode when > user manually starts heterogeneous task executors. > We need to calculate the registered/free resources according to the exact > {{defaultSlotResourceProfile}} and {{totalResourceProfile}} of > \{{TaskExecutorRegistration}}s. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] godfreyhe commented on a change in pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
godfreyhe commented on a change in pull request #14699: URL: https://github.com/apache/flink/pull/14699#discussion_r560654175 ## File path: flink-table/flink-table-planner-blink/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecIntervalJoin.java ## @@ -0,0 +1,352 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.table.planner.plan.nodes.exec.stream; + +import org.apache.flink.api.common.functions.FlatMapFunction; +import org.apache.flink.api.common.functions.MapFunction; +import org.apache.flink.api.common.typeinfo.TypeInformation; +import org.apache.flink.api.dag.Transformation; +import org.apache.flink.api.java.typeutils.ResultTypeQueryable; +import org.apache.flink.streaming.api.operators.StreamFlatMap; +import org.apache.flink.streaming.api.operators.StreamMap; +import org.apache.flink.streaming.api.operators.co.KeyedCoProcessOperator; +import org.apache.flink.streaming.api.transformations.OneInputTransformation; +import org.apache.flink.streaming.api.transformations.TwoInputTransformation; +import org.apache.flink.streaming.api.transformations.UnionTransformation; +import org.apache.flink.table.api.TableException; +import org.apache.flink.table.data.RowData; +import org.apache.flink.table.planner.delegation.PlannerBase; +import org.apache.flink.table.planner.plan.nodes.exec.ExecEdge; +import org.apache.flink.table.planner.plan.nodes.exec.ExecNode; +import org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase; +import org.apache.flink.table.planner.plan.nodes.exec.utils.JoinSpec; +import org.apache.flink.table.planner.plan.utils.IntervalJoinUtil; +import org.apache.flink.table.planner.plan.utils.JoinUtil; +import org.apache.flink.table.planner.plan.utils.KeySelectorUtil; +import org.apache.flink.table.runtime.generated.GeneratedJoinCondition; +import org.apache.flink.table.runtime.keyselector.RowDataKeySelector; +import org.apache.flink.table.runtime.operators.join.KeyedCoProcessOperatorWithWatermarkDelay; +import org.apache.flink.table.runtime.operators.join.OuterJoinPaddingUtil; +import org.apache.flink.table.runtime.operators.join.interval.IntervalJoinFunction; +import org.apache.flink.table.runtime.operators.join.interval.ProcTimeIntervalJoin; +import org.apache.flink.table.runtime.operators.join.interval.RowTimeIntervalJoin; +import org.apache.flink.table.runtime.typeutils.InternalTypeInfo; +import org.apache.flink.table.types.logical.RowType; +import org.apache.flink.util.Collector; + +import org.apache.flink.shaded.guava18.com.google.common.collect.Lists; + +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +/** {@link StreamExecNode} for a time interval stream join. */ +public class StreamExecIntervalJoin extends ExecNodeBase +implements StreamExecNode { + +private static final Logger LOGGER = LoggerFactory.getLogger(StreamExecIntervalJoin.class); + +private final JoinSpec joinSpec; +private final IntervalJoinUtil.WindowBounds windowBounds; + +public StreamExecIntervalJoin( +JoinSpec joinSpec, +IntervalJoinUtil.WindowBounds windowBounds, Review comment: port `WindowBounds` to java ? ## File path: flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/operators/join/interval/IntervalJoinFunction.java ## @@ -0,0 +1,75 @@ +package org.apache.flink.table.runtime.operators.join.interval; + +import org.apache.flink.api.common.functions.RichFlatJoinFunction; +import org.apache.flink.api.common.typeinfo.TypeInformation; +import org.apache.flink.api.java.typeutils.ResultTypeQueryable; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.table.data.RowData; +import org.apache.flink.table.data.binary.BinaryRowData; +import org.apache.flink.table.data.utils.JoinedRowData; +import org.apache.flink.table.runtime.generated.GeneratedJoinCondition; +import org.apache.flink.table.runtime.generated.JoinCondition; +import org.apache.flink.table.runtime.operators.join.NullAwareJoinHelper; +import org.apache.flink.table.runtime.typeutils.InternalTypeInfo; +import org.apache.flink.util.Collector; + +/** + * {@link
[jira] [Created] (FLINK-21047) Expose the correct registered/free resources information in SlotManager
Yangze Guo created FLINK-21047: -- Summary: Expose the correct registered/free resources information in SlotManager Key: FLINK-21047 URL: https://issues.apache.org/jira/browse/FLINK-21047 Project: Flink Issue Type: Improvement Reporter: Yangze Guo In FLINK-16640, we extend ResourceOverview and TaskManager(Details)Info for registered/free resources. However, the implementation is based on the assumption that all the registered task executors are homogeneous with the default resource spec. This assumption will be broken in standalone mode when user manually starts heterogeneous task executors. We need to calculate the registered/free resources according to the exact {{defaultSlotResourceProfile}} and {{totalResourceProfile}} of \{{TaskExecutorRegistration}}s. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #14704: [FLINK-21044][connectors/jdbc] Use more random bytes in Xid
flinkbot commented on pull request #14704: URL: https://github.com/apache/flink/pull/14704#issuecomment-763375645 ## CI report: * d37ad7dc90d2f07f30b9db9fc5185d591f1189f8 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14703: [FLINK-21039][doc] change legacy_planner.zh.md some English links to Zh links
flinkbot edited a comment on pull request #14703: URL: https://github.com/apache/flink/pull/14703#issuecomment-763364170 ## CI report: * 01f6d4f1539b32d82345da6240afde6a10653ac6 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12265) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14659: [FLINK-20931][coordination] Remove globalModVersion from ExecutionGraph
flinkbot edited a comment on pull request #14659: URL: https://github.com/apache/flink/pull/14659#issuecomment-760700709 ## CI report: * 21f45aeedd834992a8cafa3e8d39ad230f943cf3 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12244) * f9e6021770ed79f0dd9aa7df316dc6a9b4f4ae0e Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12264) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-docker] Myasuka commented on a change in pull request #57: [FLINK-21034] Refactor jemalloc switch to environment variable
Myasuka commented on a change in pull request #57: URL: https://github.com/apache/flink-docker/pull/57#discussion_r560710585 ## File path: testing/bin/config.sh ## @@ -0,0 +1,6 @@ +#!/usr/bin/env bash Review comment: This file is missing apache license. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Comment Edited] (FLINK-20946) Optimize Python ValueState Implementation In PyFlink
[ https://issues.apache.org/jira/browse/FLINK-20946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268394#comment-17268394 ] Dian Fu edited comment on FLINK-20946 at 1/20/21, 6:38 AM: --- Reverted in a1588e55d2a88282882d686478218807f6332691 as it causes test instability issues. See FLINK-21046 for more details was (Author: dian.fu): Reverted in a1588e55d2a88282882d686478218807f6332691 as it causes test instability issues. > Optimize Python ValueState Implementation In PyFlink > > > Key: FLINK-20946 > URL: https://issues.apache.org/jira/browse/FLINK-20946 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: Huang Xingbo >Priority: Major > Labels: pull-request-available > Fix For: 1.13.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Closed] (FLINK-21046) test_map_view and test_map_view_iterate test failed
[ https://issues.apache.org/jira/browse/FLINK-21046?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dian Fu closed FLINK-21046. --- Resolution: Fixed [~hxbks2ks] I'm closing this ticket. Let's fix this problem in FLINK-20946 > test_map_view and test_map_view_iterate test failed > --- > > Key: FLINK-21046 > URL: https://issues.apache.org/jira/browse/FLINK-21046 > Project: Flink > Issue Type: Bug > Components: API / Python >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12257=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=8d78fe4f-d658-5c70-12f8-4921589024c3] > {code:java} > 2021-01-20T02:25:30.7107189Z E py4j.protocol.Py4JJavaError: > An error occurred while calling > z:org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame. > 2021-01-20T02:25:30.7108168Z E : > java.lang.RuntimeException: Could not remove element '+I[Hi,Hi2,Hi3,Hi_, > 1,1,1,2, Hi:2,Hi2:1,Hi3:1,Hi_:1, 3, hi]', should never happen. > 2021-01-20T02:25:30.7108821Z Eat > org.apache.flink.table.runtime.arrow.ArrowUtils.filterOutRetractRows(ArrowUtils.java:754) > 2021-01-20T02:25:30.7109502Z Eat > org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame(ArrowUtils.java:673) > 2021-01-20T02:25:30.7110005Z Eat > sun.reflect.GeneratedMethodAccessor264.invoke(Unknown Source) > 2021-01-20T02:25:30.7110481Z Eat > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2021-01-20T02:25:30.7110983Z Eat > java.lang.reflect.Method.invoke(Method.java:498) > 2021-01-20T02:25:30.7111503Z Eat > org.apache.flink.api.python.shaded.py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) > 2021-01-20T02:25:30.7112281Z Eat > org.apache.flink.api.python.shaded.py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) > 2021-01-20T02:25:30.7113163Z Eat > org.apache.flink.api.python.shaded.py4j.Gateway.invoke(Gateway.java:282) > 2021-01-20T02:25:30.7114101Z Eat > org.apache.flink.api.python.shaded.py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) > 2021-01-20T02:25:30.7115014Z Eat > org.apache.flink.api.python.shaded.py4j.commands.CallCommand.execute(CallCommand.java:79) > 2021-01-20T02:25:30.7115884Z Eat > org.apache.flink.api.python.shaded.py4j.GatewayConnection.run(GatewayConnection.java:238) > 2021-01-20T02:25:30.7116379Z Eat > java.lang.Thread.run(Thread.java:748) > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Reopened] (FLINK-20946) Optimize Python ValueState Implementation In PyFlink
[ https://issues.apache.org/jira/browse/FLINK-20946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dian Fu reopened FLINK-20946: - Reverted in a1588e55d2a88282882d686478218807f6332691 as it causes test instability issues. > Optimize Python ValueState Implementation In PyFlink > > > Key: FLINK-20946 > URL: https://issues.apache.org/jira/browse/FLINK-20946 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: Huang Xingbo >Priority: Major > Labels: pull-request-available > Fix For: 1.13.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-21046) test_map_view and test_map_view_iterate test failed
[ https://issues.apache.org/jira/browse/FLINK-21046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268393#comment-17268393 ] Dian Fu commented on FLINK-21046: - Thanks [~hxbks2ks]. It seems caused by FLINK-20946. I have reverted it as a quick fix~ > test_map_view and test_map_view_iterate test failed > --- > > Key: FLINK-21046 > URL: https://issues.apache.org/jira/browse/FLINK-21046 > Project: Flink > Issue Type: Bug > Components: API / Python >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12257=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=8d78fe4f-d658-5c70-12f8-4921589024c3] > {code:java} > 2021-01-20T02:25:30.7107189Z E py4j.protocol.Py4JJavaError: > An error occurred while calling > z:org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame. > 2021-01-20T02:25:30.7108168Z E : > java.lang.RuntimeException: Could not remove element '+I[Hi,Hi2,Hi3,Hi_, > 1,1,1,2, Hi:2,Hi2:1,Hi3:1,Hi_:1, 3, hi]', should never happen. > 2021-01-20T02:25:30.7108821Z Eat > org.apache.flink.table.runtime.arrow.ArrowUtils.filterOutRetractRows(ArrowUtils.java:754) > 2021-01-20T02:25:30.7109502Z Eat > org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame(ArrowUtils.java:673) > 2021-01-20T02:25:30.7110005Z Eat > sun.reflect.GeneratedMethodAccessor264.invoke(Unknown Source) > 2021-01-20T02:25:30.7110481Z Eat > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2021-01-20T02:25:30.7110983Z Eat > java.lang.reflect.Method.invoke(Method.java:498) > 2021-01-20T02:25:30.7111503Z Eat > org.apache.flink.api.python.shaded.py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) > 2021-01-20T02:25:30.7112281Z Eat > org.apache.flink.api.python.shaded.py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) > 2021-01-20T02:25:30.7113163Z Eat > org.apache.flink.api.python.shaded.py4j.Gateway.invoke(Gateway.java:282) > 2021-01-20T02:25:30.7114101Z Eat > org.apache.flink.api.python.shaded.py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) > 2021-01-20T02:25:30.7115014Z Eat > org.apache.flink.api.python.shaded.py4j.commands.CallCommand.execute(CallCommand.java:79) > 2021-01-20T02:25:30.7115884Z Eat > org.apache.flink.api.python.shaded.py4j.GatewayConnection.run(GatewayConnection.java:238) > 2021-01-20T02:25:30.7116379Z Eat > java.lang.Thread.run(Thread.java:748) > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #14704: [FLINK-21044][connectors/jdbc] Use more random bytes in Xid
flinkbot commented on pull request #14704: URL: https://github.com/apache/flink/pull/14704#issuecomment-763368940 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit d37ad7dc90d2f07f30b9db9fc5185d591f1189f8 (Wed Jan 20 06:25:56 UTC 2021) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-21044) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
[ https://issues.apache.org/jira/browse/FLINK-21044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-21044: --- Labels: pull-request-available test-stability (was: test-stability) > SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed > -- > > Key: FLINK-21044 > URL: https://issues.apache.org/jira/browse/FLINK-21044 > Project: Flink > Issue Type: Bug > Components: Connectors / JDBC >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: Roman Khachatryan >Priority: Critical > Labels: pull-request-available, test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] > {code:java} > 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, > Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in > org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest > 2021-01-19T15:21:35.5665936Z [ERROR] > testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) > Time elapsed: 0.024 s <<< FAILURE! > 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: > expected:<1> but was:<> > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] rkhachatryan opened a new pull request #14704: [FLINK-21044][connectors/jdbc] Use more random bytes in Xid
rkhachatryan opened a new pull request #14704: URL: https://github.com/apache/flink/pull/14704 ## What is the purpose of the change With only 4 random bytes, `testXidsUniqueAmongGenerators` fails once per ~100 runs as it deliberately uses same checkpointId and subtaskIdx. This change makes Semantic`XidGenerator` use 8 random bytes instead of 4 higher bytes of `checkpointId`. ## Verifying this change Already covered by `testXidsUniqueAmongGenerators`. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: no - The serializers: no - The runtime per-record code paths (performance sensitive): no - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? no - If yes, how is the feature documented? no This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #14703: [FLINK-21039][doc] change legacy_planner.zh.md some English links to Zh links
flinkbot commented on pull request #14703: URL: https://github.com/apache/flink/pull/14703#issuecomment-763364170 ## CI report: * 01f6d4f1539b32d82345da6240afde6a10653ac6 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14659: [FLINK-20931][coordination] Remove globalModVersion from ExecutionGraph
flinkbot edited a comment on pull request #14659: URL: https://github.com/apache/flink/pull/14659#issuecomment-760700709 ## CI report: * 21f45aeedd834992a8cafa3e8d39ad230f943cf3 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12244) * f9e6021770ed79f0dd9aa7df316dc6a9b4f4ae0e UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #14703: [FLINK-21039][doc] change legacy_planner.zh.md some English links to Zh links
flinkbot commented on pull request #14703: URL: https://github.com/apache/flink/pull/14703#issuecomment-763360264 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 01f6d4f1539b32d82345da6240afde6a10653ac6 (Wed Jan 20 06:07:20 UTC 2021) ✅no warnings Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-21046) test_map_view and test_map_view_iterate test failed
[ https://issues.apache.org/jira/browse/FLINK-21046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268379#comment-17268379 ] Huang Xingbo commented on FLINK-21046: -- I wil fix it asap > test_map_view and test_map_view_iterate test failed > --- > > Key: FLINK-21046 > URL: https://issues.apache.org/jira/browse/FLINK-21046 > Project: Flink > Issue Type: Bug > Components: API / Python >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12257=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=8d78fe4f-d658-5c70-12f8-4921589024c3] > {code:java} > 2021-01-20T02:25:30.7107189Z E py4j.protocol.Py4JJavaError: > An error occurred while calling > z:org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame. > 2021-01-20T02:25:30.7108168Z E : > java.lang.RuntimeException: Could not remove element '+I[Hi,Hi2,Hi3,Hi_, > 1,1,1,2, Hi:2,Hi2:1,Hi3:1,Hi_:1, 3, hi]', should never happen. > 2021-01-20T02:25:30.7108821Z Eat > org.apache.flink.table.runtime.arrow.ArrowUtils.filterOutRetractRows(ArrowUtils.java:754) > 2021-01-20T02:25:30.7109502Z Eat > org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame(ArrowUtils.java:673) > 2021-01-20T02:25:30.7110005Z Eat > sun.reflect.GeneratedMethodAccessor264.invoke(Unknown Source) > 2021-01-20T02:25:30.7110481Z Eat > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2021-01-20T02:25:30.7110983Z Eat > java.lang.reflect.Method.invoke(Method.java:498) > 2021-01-20T02:25:30.7111503Z Eat > org.apache.flink.api.python.shaded.py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) > 2021-01-20T02:25:30.7112281Z Eat > org.apache.flink.api.python.shaded.py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) > 2021-01-20T02:25:30.7113163Z Eat > org.apache.flink.api.python.shaded.py4j.Gateway.invoke(Gateway.java:282) > 2021-01-20T02:25:30.7114101Z Eat > org.apache.flink.api.python.shaded.py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) > 2021-01-20T02:25:30.7115014Z Eat > org.apache.flink.api.python.shaded.py4j.commands.CallCommand.execute(CallCommand.java:79) > 2021-01-20T02:25:30.7115884Z Eat > org.apache.flink.api.python.shaded.py4j.GatewayConnection.run(GatewayConnection.java:238) > 2021-01-20T02:25:30.7116379Z Eat > java.lang.Thread.run(Thread.java:748) > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21039) Broken links in "dev/table/legacy_planner.zh.md"
[ https://issues.apache.org/jira/browse/FLINK-21039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-21039: --- Labels: pull-request-available (was: ) > Broken links in "dev/table/legacy_planner.zh.md" > > > Key: FLINK-21039 > URL: https://issues.apache.org/jira/browse/FLINK-21039 > Project: Flink > Issue Type: Bug > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: WeiNan Zhao >Priority: Critical > Labels: pull-request-available > > dev/table/legacy_planner.zh.md contains some English links causing errors. > {code:java} > Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. > Make sure the document exists and the path is correct. in > dev/table/legacy_planner.zh.md > Could not find document 'dev/batch/index.md' in tag 'link'. > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-21046) test_map_view and test_map_view_iterate test failed
Huang Xingbo created FLINK-21046: Summary: test_map_view and test_map_view_iterate test failed Key: FLINK-21046 URL: https://issues.apache.org/jira/browse/FLINK-21046 Project: Flink Issue Type: Bug Components: API / Python Affects Versions: 1.13.0 Reporter: Huang Xingbo [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12257=logs=9cada3cb-c1d3-5621-16da-0f718fb86602=8d78fe4f-d658-5c70-12f8-4921589024c3] {code:java} 2021-01-20T02:25:30.7107189Z E py4j.protocol.Py4JJavaError: An error occurred while calling z:org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame. 2021-01-20T02:25:30.7108168Z E : java.lang.RuntimeException: Could not remove element '+I[Hi,Hi2,Hi3,Hi_, 1,1,1,2, Hi:2,Hi2:1,Hi3:1,Hi_:1, 3, hi]', should never happen. 2021-01-20T02:25:30.7108821Z E at org.apache.flink.table.runtime.arrow.ArrowUtils.filterOutRetractRows(ArrowUtils.java:754) 2021-01-20T02:25:30.7109502Z E at org.apache.flink.table.runtime.arrow.ArrowUtils.collectAsPandasDataFrame(ArrowUtils.java:673) 2021-01-20T02:25:30.7110005Z E at sun.reflect.GeneratedMethodAccessor264.invoke(Unknown Source) 2021-01-20T02:25:30.7110481Z E at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 2021-01-20T02:25:30.7110983Z E at java.lang.reflect.Method.invoke(Method.java:498) 2021-01-20T02:25:30.7111503Z E at org.apache.flink.api.python.shaded.py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) 2021-01-20T02:25:30.7112281Z E at org.apache.flink.api.python.shaded.py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) 2021-01-20T02:25:30.7113163Z E at org.apache.flink.api.python.shaded.py4j.Gateway.invoke(Gateway.java:282) 2021-01-20T02:25:30.7114101Z E at org.apache.flink.api.python.shaded.py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) 2021-01-20T02:25:30.7115014Z E at org.apache.flink.api.python.shaded.py4j.commands.CallCommand.execute(CallCommand.java:79) 2021-01-20T02:25:30.7115884Z E at org.apache.flink.api.python.shaded.py4j.GatewayConnection.run(GatewayConnection.java:238) 2021-01-20T02:25:30.7116379Z E at java.lang.Thread.run(Thread.java:748) {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] chaozwn opened a new pull request #14703: [FLINK-21039][doc] change legacy_planner.zh.md some English links to Zh links
chaozwn opened a new pull request #14703: URL: https://github.com/apache/flink/pull/14703 ## What is the purpose of the change change legacy_planner.zh.md some English links to Zh links ## Brief change log change legacy_planner.zh.md some English links to Zh links ## Verifying this change This change is already covered by existing tests, such as *(please describe tests)*. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: no - The serializers: no - The runtime per-record code paths (performance sensitive): no - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? no - If yes, how is the feature documented? docs This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-21045) Support 'load module' and 'unload module' SQL syntax
Nicholas Jiang created FLINK-21045: -- Summary: Support 'load module' and 'unload module' SQL syntax Key: FLINK-21045 URL: https://issues.apache.org/jira/browse/FLINK-21045 Project: Flink Issue Type: Improvement Components: Table SQL / Planner Affects Versions: 1.13.0 Reporter: Nicholas Jiang At present, Flink SQL doesn't support the 'load module' and 'unload module' SQL syntax. It's necessary for uses in the situation that users load and unload user-defined module through table api or sql client. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Closed] (FLINK-21043) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
[ https://issues.apache.org/jira/browse/FLINK-21043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu closed FLINK-21043. --- Resolution: Duplicate > SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed > -- > > Key: FLINK-21043 > URL: https://issues.apache.org/jira/browse/FLINK-21043 > Project: Flink > Issue Type: Bug > Components: Connectors / JDBC >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] > {code:java} > 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, > Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in > org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest > 2021-01-19T15:21:35.5665936Z [ERROR] > testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) > Time elapsed: 0.024 s <<< FAILURE! > 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: > expected:<1> but was:<> > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Assigned] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu reassigned FLINK-21042: --- Assignee: Roc Marshal > Correct the error in the code example in section 'aggregate-functions' of > table-page 'udfs'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Assignee: Roc Marshal >Priority: Major > Labels: pull-request-available > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #14702: [FLINK-21042][docs] Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
flinkbot edited a comment on pull request #14702: URL: https://github.com/apache/flink/pull/14702#issuecomment-763339190 ## CI report: * cf08ee38181b6c095fe7673197c90b9eabeb450d Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12263) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-21044) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
[ https://issues.apache.org/jira/browse/FLINK-21044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268367#comment-17268367 ] Roman Khachatryan commented on FLINK-21044: --- Sure, I'll take a look. > SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed > -- > > Key: FLINK-21044 > URL: https://issues.apache.org/jira/browse/FLINK-21044 > Project: Flink > Issue Type: Bug > Components: Connectors / JDBC >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: Roman Khachatryan >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] > {code:java} > 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, > Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in > org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest > 2021-01-19T15:21:35.5665936Z [ERROR] > testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) > Time elapsed: 0.024 s <<< FAILURE! > 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: > expected:<1> but was:<> > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #14702: [FLINK-21042][docs] Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
flinkbot commented on pull request #14702: URL: https://github.com/apache/flink/pull/14702#issuecomment-763339190 ## CI report: * cf08ee38181b6c095fe7673197c90b9eabeb450d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-21044) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
[ https://issues.apache.org/jira/browse/FLINK-21044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Roman Khachatryan reassigned FLINK-21044: - Assignee: Roman Khachatryan > SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed > -- > > Key: FLINK-21044 > URL: https://issues.apache.org/jira/browse/FLINK-21044 > Project: Flink > Issue Type: Bug > Components: Connectors / JDBC >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: Roman Khachatryan >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] > {code:java} > 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, > Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in > org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest > 2021-01-19T15:21:35.5665936Z [ERROR] > testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) > Time elapsed: 0.024 s <<< FAILURE! > 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: > expected:<1> but was:<> > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-21044) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
[ https://issues.apache.org/jira/browse/FLINK-21044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268366#comment-17268366 ] Huang Xingbo commented on FLINK-21044: -- cc [~roman_khachatryan] Could you help take a look? Thanks. > SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed > -- > > Key: FLINK-21044 > URL: https://issues.apache.org/jira/browse/FLINK-21044 > Project: Flink > Issue Type: Bug > Components: Connectors / JDBC >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] > {code:java} > 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, > Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in > org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest > 2021-01-19T15:21:35.5665936Z [ERROR] > testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) > Time elapsed: 0.024 s <<< FAILURE! > 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: > expected:<1> but was:<> > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-21044) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
Huang Xingbo created FLINK-21044: Summary: SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed Key: FLINK-21044 URL: https://issues.apache.org/jira/browse/FLINK-21044 Project: Flink Issue Type: Bug Components: Connectors / JDBC Affects Versions: 1.13.0 Reporter: Huang Xingbo [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] {code:java} 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest 2021-01-19T15:21:35.5665936Z [ERROR] testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) Time elapsed: 0.024 s <<< FAILURE! 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: expected:<1> but was:<> {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-21043) SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed
Huang Xingbo created FLINK-21043: Summary: SemanticXidGeneratorTest.testXidsUniqueAmongGenerators test failed Key: FLINK-21043 URL: https://issues.apache.org/jira/browse/FLINK-21043 Project: Flink Issue Type: Bug Components: Connectors / JDBC Affects Versions: 1.13.0 Reporter: Huang Xingbo [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12245=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20] {code:java} 2021-01-19T15:21:35.5665063Z [ERROR] Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.1 s <<< FAILURE! - in org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest 2021-01-19T15:21:35.5665936Z [ERROR] testXidsUniqueAmongGenerators(org.apache.flink.connector.jdbc.xa.SemanticXidGeneratorTest) Time elapsed: 0.024 s <<< FAILURE! 2021-01-19T15:21:35.5666770Z junit.framework.AssertionFailedError: expected:<1> but was:<> {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #14702: [FLINK-21042][docs] Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
flinkbot commented on pull request #14702: URL: https://github.com/apache/flink/pull/14702#issuecomment-76920 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit cf08ee38181b6c095fe7673197c90b9eabeb450d (Wed Jan 20 05:03:17 UTC 2021) **Warnings:** * Documentation files were touched, but no `.zh.md` files: Update Chinese documentation or file Jira ticket. * **This pull request references an unassigned [Jira ticket](https://issues.apache.org/jira/browse/FLINK-21042).** According to the [code contribution guide](https://flink.apache.org/contributing/contribute-code.html), tickets need to be assigned before starting with the implementation work. Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-21042: --- Labels: pull-request-available (was: ) > Correct the error in the code example in section 'aggregate-functions' of > table-page 'udfs'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: pull-request-available > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] RocMarshal opened a new pull request #14702: [FLINK-21042][docs] Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
RocMarshal opened a new pull request #14702: URL: https://github.com/apache/flink/pull/14702 ## What is the purpose of the change *Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.* The markdown file location: flink/docs/dev/table/functions/udfs.md The target url : https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions ## Brief change log *Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.* - The 'value' is a reserved keyword of the flink table module,So the correct usage should to execute in back quotes .* ## Verifying this change *Pure documentation change.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / **no**) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / **no**) - The serializers: (yes / **no** / don't know) - The runtime per-record code paths (performance sensitive): (yes / **no** / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (yes / **no** / don't know) - The S3 file system connector: (yes / **no** / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / **no**) - If yes, how is the feature documented? (**not applicable** / docs / JavaDocs / not documented) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] anlen321 commented on pull request #14684: [FLINK-20460][Connector-HBase] Support async lookup for HBase connector
anlen321 commented on pull request #14684: URL: https://github.com/apache/flink/pull/14684#issuecomment-763327328 @leonardBang My current approach to conflict is not friendly and resulted in the inclusion of other commits. So I want to create a new branch and apply for a new pull request. Do you think it is OK? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14701: [FLINK-20500][upsert-kafka] Fix temporal join test
flinkbot edited a comment on pull request #14701: URL: https://github.com/apache/flink/pull/14701#issuecomment-763319611 ## CI report: * bd324d7ada655e64f14da42e4f6e5e0ea94e7751 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12262) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] leonardBang commented on a change in pull request #14675: [FLINK-18383][docs]translate "JDBC SQL Connector" documentation into …
leonardBang commented on a change in pull request #14675: URL: https://github.com/apache/flink/pull/14675#discussion_r560663060 ## File path: docs/dev/table/connectors/jdbc.zh.md ## @@ -31,35 +31,33 @@ under the License. * This will be replaced by the TOC {:toc} -The JDBC connector allows for reading data from and writing data into any relational databases with a JDBC driver. This document describes how to setup the JDBC connector to run SQL queries against relational databases. +JDBC连接器允许使用JDBC驱动向任意类型的关系型数据库读取或者写入数据。本文档描述了针对关系型数据库如何通过建立JDBC连接器来执行SQL查询。 -The JDBC sink operate in upsert mode for exchange UPDATE/DELETE messages with the external system if a primary key is defined on the DDL, otherwise, it operates in append mode and doesn't support to consume UPDATE/DELETE messages. +如果一个主键定义在 DDL 中,JDBC sink 将以 upsert 模式与外部系统交换 UPDATE/DELETE 消息;否则,它将以appened 模式与外部系统交换消息且不支持消费 UPDATE/DELETE 消息。 -Dependencies + +依赖 {% assign connector = site.data.sql-connectors['jdbc'] %} -{% include sql-connector-download-table.zh.html -connector=connector -%} +{% include sql-connector-download-table.zh.html connector='connector' %} -A driver dependency is also required to connect to a specified database. Here are drivers currently supported: +在连接到具体数据库时,也需要对应的驱动依赖,目前支持的驱动如下: | Driver | Group Id | Artifact Id | JAR | | :---| :--| :--| :| -| MySQL | `mysql` | `mysql-connector-java` | [Download](https://repo.maven.apache.org/maven2/mysql/mysql-connector-java/) | -| PostgreSQL | `org.postgresql` | `postgresql` | [Download](https://jdbc.postgresql.org/download.html) | -| Derby | `org.apache.derby` |`derby` | [Download](http://db.apache.org/derby/derby_downloads.html) | +| MySQL | `mysql` | `mysql-connector-java` | [下载](https://repo.maven.apache.org/maven2/mysql/mysql-connector-java/) | +| PostgreSQL | `org.postgresql` | `postgresql` | [下载](https://jdbc.postgresql.org/download.html) | +| Derby | `org.apache.derby` |`derby` | [下载](http://db.apache.org/derby/derby_downloads.html) | -JDBC connector and drivers are not currently part of Flink's binary distribution. See how to link with them for cluster execution [here]({% link dev/project-configuration.zh.md %}). - +当前,JDBC 连接器和驱动不在 Flink 二进制发布包中,请参阅 [这里]({% link dev/project-configuration.zh.md %}) 了解在集群上执行时何连接它们。 -How to create a JDBC table +如何创建 JDBC table Review comment: ```suggestion 如何创建 JDBC 表 ``` ## File path: docs/dev/table/connectors/jdbc.zh.md ## @@ -92,210 +90,206 @@ ON myTopic.key = MyUserTable.id; -Connector Options +连接器参数选项 Review comment: ```suggestion 连接器参数 ``` ## File path: docs/dev/table/connectors/jdbc.zh.md ## @@ -92,210 +90,206 @@ ON myTopic.key = MyUserTable.id; -Connector Options +连接器参数选项 -Option -Required -Default -Type -Description +选项 Review comment: 和其他中文文档保持一致吧 ## File path: docs/dev/table/connectors/jdbc.zh.md ## @@ -92,210 +90,206 @@ ON myTopic.key = MyUserTable.id; -Connector Options +连接器参数选项 -Option -Required -Default -Type -Description +选项 Review comment: ```suggestion 参数 ``` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #14701: [FLINK-20500][upsert-kafka] Fix temporal join test
flinkbot commented on pull request #14701: URL: https://github.com/apache/flink/pull/14701#issuecomment-763319611 ## CI report: * bd324d7ada655e64f14da42e4f6e5e0ea94e7751 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14700: [FLINK-21041][table-planner-blink] Introduce ExecNodeGraph to wrap the ExecNode topology
flinkbot edited a comment on pull request #14700: URL: https://github.com/apache/flink/pull/14700#issuecomment-763313006 ## CI report: * f60392a83e02d26e7a25c2100c3fd85fac4f54e2 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12261) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-21042: Component/s: Table SQL / Ecosystem > Correct the error in the code example in section 'aggregate-functions' of > table-page 'udfs'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: Document, Documentation, docuentation > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268357#comment-17268357 ] Roc Marshal edited comment on FLINK-21042 at 1/20/21, 4:11 AM: --- [~jark] [~appleyuchi] Could you assign this ticket to me ? Thank you very much! was (Author: rocmarshal): [~jark] Could you assign this ticket to me ? Thank you very much! > Correct the error in the code example in section 'aggregate-functions' of > table-page 'udfs'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: Document, Documentation, docuentation > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-21042: Labels: (was: Document Documentation docuentation) > Correct the error in the code example in section 'aggregate-functions' of > table-page 'udfs'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Roc Marshal updated FLINK-21042: Summary: Correct the error in the code example in section 'aggregate-functions' of table-page 'udfs'. (was: Correct the error in the code example in section 'aggregate-functions' of page 'udfs'.) > Correct the error in the code example in section 'aggregate-functions' of > table-page 'udfs'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: Document, Documentation, docuentation > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21042) Correct the error in the code example in section 'aggregate-functions' of page 'udfs'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Roc Marshal updated FLINK-21042: Summary: Correct the error in the code example in section 'aggregate-functions' of page 'udfs'. (was: Correct the error in the code example in page 'aggregate-functions'.) > Correct the error in the code example in section 'aggregate-functions' of > page 'udfs'. > -- > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: Document, Documentation, docuentation > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21042) Correct the error in the code example in page 'aggregate-functions'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Roc Marshal updated FLINK-21042: Description: The markdown file location: flink/docs/dev/table/functions/udfs.md The target url : [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] The original content: *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* The 'value' is a reserved keyword of the flink table module,So the correct usage should to execute in back quotes . _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ was: The markdown file location: flink/docs/dev/table/functions/udfs.zh.md & flink/docs/dev/table/functions/udfs.md The target url : [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] The original content: *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* The 'value' is a reserved keyword of the flink table module,So the correct usage should to execute in back quotes . _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ > Correct the error in the code example in page 'aggregate-functions'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: Document, Documentation, docuentation > > The markdown file location: flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-21042) Correct the error in the code example in page 'aggregate-functions'.
[ https://issues.apache.org/jira/browse/FLINK-21042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268357#comment-17268357 ] Roc Marshal commented on FLINK-21042: - [~jark] Could you assign this ticket to me ? Thank you very much! > Correct the error in the code example in page 'aggregate-functions'. > > > Key: FLINK-21042 > URL: https://issues.apache.org/jira/browse/FLINK-21042 > Project: Flink > Issue Type: Improvement > Components: Documentation >Affects Versions: 1.13.0, 1.12.1 >Reporter: Roc Marshal >Priority: Major > Labels: Document, Documentation, docuentation > > The markdown file location: flink/docs/dev/table/functions/udfs.zh.md & > flink/docs/dev/table/functions/udfs.md > The target url : > [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] > > The original content: > *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* > > The 'value' is a reserved keyword of the flink table module,So the correct > usage should to execute in back quotes . > _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-21042) Correct the error in the code example in page 'aggregate-functions'.
Roc Marshal created FLINK-21042: --- Summary: Correct the error in the code example in page 'aggregate-functions'. Key: FLINK-21042 URL: https://issues.apache.org/jira/browse/FLINK-21042 Project: Flink Issue Type: Improvement Components: Documentation Affects Versions: 1.12.1, 1.13.0 Reporter: Roc Marshal The markdown file location: flink/docs/dev/table/functions/udfs.zh.md & flink/docs/dev/table/functions/udfs.md The target url : [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/udfs.html#aggregate-functions] The original content: *SELECT myField, WeightedAvg(value, weight) FROM MyTable GROUP BY myField* The 'value' is a reserved keyword of the flink table module,So the correct usage should to execute in back quotes . _*SELECT myField, WeightedAvg(`value`, weight) FROM MyTable GROUP BY myField*_ -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #14700: [FLINK-21041][table-planner-blink] Introduce ExecNodeGraph to wrap the ExecNode topology
flinkbot commented on pull request #14700: URL: https://github.com/apache/flink/pull/14700#issuecomment-763313006 ## CI report: * f60392a83e02d26e7a25c2100c3fd85fac4f54e2 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14697: [FLINK-21024][dist] Keep dynamic properties before job arguments.
flinkbot edited a comment on pull request #14697: URL: https://github.com/apache/flink/pull/14697#issuecomment-762786872 ## CI report: * 2bcacd1650a73b29652336934d514b423f3f636a Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12237) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #14701: [FLINK-20500][upsert-kafka] Fix temporal join test
flinkbot commented on pull request #14701: URL: https://github.com/apache/flink/pull/14701#issuecomment-763312576 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit bd324d7ada655e64f14da42e4f6e5e0ea94e7751 (Wed Jan 20 03:57:13 UTC 2021) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] fsk119 opened a new pull request #14701: [FLINK-20500][upsert-kafka] Fix temporal join test
fsk119 opened a new pull request #14701: URL: https://github.com/apache/flink/pull/14701 ## What is the purpose of the change *The reason why the test fails is the records in the same partition are out-of-order. I print the partition-id and offset in the partition.* ``` partition-0 <102,Greg,Berlin,BERLIN,2020-08-15T00:00:03,0,0> <103,Richard,Berlin,BERLIN,2020-08-16T00:01:05,0,1> <103,Richard,London,LONDON,2020-08-16T01:01:05,0,2> <105,Tim,Shenzhen,SHENZHEN,2020-08-16T00:06,0,3> <105,Tim,Hongkong,HONGKONG,2020-08-16T01:06,0,4> partition-1 <100,Bob,Beijing,BEIJING,2020-08-15T00:00:01,1,0> <101,Alice,Shanghai,SHANGHAI,2020-08-15T00:00:02,1,1> <101,Alice,Wuhan,WUHAN,2020-08-16T00:02,1,2> <101,Alice,Hangzhou,HANGZHOU,2020-08-16T01:04:05,1,3> <104,Tomato,Hongkong,HONGKONG,2020-08-16T00:05:05,1,4> <104,Tomato,Shenzhen,SHENZHEN,2020-08-16T01:05:05,1,5> ``` In the test, the expected record is `<10004,104,2020-08-16T00:05:06,Tomato,HONGKONG,2020-08-16T00:05:05> ` but the actual record is `<10004,104,2020-08-16T00:05:06,null,null,null>`. It may be the left stream record `<10004,104,2020-08-16T00:05:06>` arrives before the record `<104,Tomato,Hongkong,HONGKONG,2020-08-16T00:05:05,1,4> ` in the right stream. When the left stream arrives, it finds the watermark is at timestamp `2020-08-16T01:04:05` in the right stream, which means the records before `2020-08-16T01:04:05` have arrived and emits itself. Therefore, we should adjust the order in the partition or watermark strategy. Here we just adjust order for convenience. The new order in the partition follows. ``` <102,Greg,Berlin,BERLIN,2020-08-15T00:00:03,0,0> <103,Richard,Berlin,BERLIN,2020-08-16T00:01:05,0,1> <105,Tim,Shenzhen,SHENZHEN,2020-08-16T00:06,0,2> <103,Richard,London,LONDON,2020-08-16T01:01:05,0,3> <105,Tim,Hongkong,HONGKONG,2020-08-16T01:06,0,4> <100,Bob,Beijing,BEIJING,2020-08-15T00:00:01,1,0> <101,Alice,Shanghai,SHANGHAI,2020-08-15T00:00:02,1,1> <101,Alice,Wuhan,WUHAN,2020-08-16T00:02,1,2> <104,Tomato,Hongkong,HONGKONG,2020-08-16T00:05:05,1,3> <101,Alice,Hangzhou,HANGZHOU,2020-08-16T01:04:05,1,4> <104,Tomato,Shenzhen,SHENZHEN,2020-08-16T01:05:05,1,5> ``` ## Brief change log - *Reorder the records in the partition* ## Documentation - Does this pull request introduce a new feature? (yes / **no**) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-13247) Implement external shuffle service for YARN
[ https://issues.apache.org/jira/browse/FLINK-13247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268352#comment-17268352 ] Zhijiang commented on FLINK-13247: -- Thanks [~trohrmann] for the operation. Actually I tried to do that while last commenting, but failed unexpectedly :( > Implement external shuffle service for YARN > --- > > Key: FLINK-13247 > URL: https://issues.apache.org/jira/browse/FLINK-13247 > Project: Flink > Issue Type: New Feature > Components: Runtime / Network >Reporter: MalcolmSanders >Priority: Minor > > Flink batch job users could achieve better cluster utilization and job > throughput throught external shuffle service because the producers of > intermedia result partitions can be released once intermedia result > partitions have been persisted on disks. In > [FLINK-10653|https://issues.apache.org/jira/browse/FLINK-10653], [~zjwang] > has introduced pluggable shuffle manager architecture which abstracts the > process of data transfer between stages from flink runtime as shuffle > service. I propose to YARN implementation for flink external shuffle service > since YARN is widely used in various companies. > The basic idea is as follows: > (1) Producers write intermedia result partitions to local disks assigned by > NodeManager; > (2) Yarn shuffle servers, deployed on each NodeManager as an auxiliary > service, are acknowledged of intermedia result partition descriptions by > producers; > (3) Consumers fetch intermedia result partition from yarn shuffle servers; -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #14697: [FLINK-21024][dist] Keep dynamic properties before job arguments.
flinkbot edited a comment on pull request #14697: URL: https://github.com/apache/flink/pull/14697#issuecomment-762786872 ## CI report: * 2bcacd1650a73b29652336934d514b423f3f636a Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12237) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14675: [FLINK-18383][docs]translate "JDBC SQL Connector" documentation into …
flinkbot edited a comment on pull request #14675: URL: https://github.com/apache/flink/pull/14675#issuecomment-761769906 ## CI report: * 7b8045005155c54e76d0e19687a1fe9309c04892 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12259) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #14700: [FLINK-21041][table-planner-blink] Introduce ExecNodeGraph to wrap the ExecNode topology
flinkbot commented on pull request #14700: URL: https://github.com/apache/flink/pull/14700#issuecomment-763304747 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit f60392a83e02d26e7a25c2100c3fd85fac4f54e2 (Wed Jan 20 03:30:04 UTC 2021) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-21041) Introduce ExecNodeGraph to wrap the ExecNode topology
[ https://issues.apache.org/jira/browse/FLINK-21041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-21041: --- Labels: pull-request-available (was: ) > Introduce ExecNodeGraph to wrap the ExecNode topology > - > > Key: FLINK-21041 > URL: https://issues.apache.org/jira/browse/FLINK-21041 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Reporter: godfrey he >Assignee: godfrey he >Priority: Major > Labels: pull-request-available > Fix For: 1.13.0 > > > Currently, we use {{List}} to represent the {{ExecNode}} > topology, as we will introduce more features (such as serialize/deserialize > {{ExecNode}}s), It's better we can introduce an unified class to represent > the topology. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] godfreyhe opened a new pull request #14700: [FLINK-21041][table-planner-blink] Introduce ExecNodeGraph to wrap the ExecNode topology
godfreyhe opened a new pull request #14700: URL: https://github.com/apache/flink/pull/14700 ## What is the purpose of the change *Currently, we use List to represent the ExecNode topology, as we will introduce more features (such as serialize/deserialize ExecNodes), It's better we can introduce an unified class to represent the topology.* ## Brief change log - *Introduce ExecNodeGraph to wrap the ExecNode topology* ## Verifying this change This change is already covered by existing tests*. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / **no**) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / **no**) - The serializers: (yes / **no** / don't know) - The runtime per-record code paths (performance sensitive): (yes / **no** / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (yes / **no** / don't know) - The S3 file system connector: (yes / **no** / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / **no**) - If yes, how is the feature documented? (not applicable / docs / JavaDocs / **not documented**) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-21041) Introduce ExecNodeGraph to wrap the ExecNode topology
[ https://issues.apache.org/jira/browse/FLINK-21041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he reassigned FLINK-21041: -- Assignee: godfrey he > Introduce ExecNodeGraph to wrap the ExecNode topology > - > > Key: FLINK-21041 > URL: https://issues.apache.org/jira/browse/FLINK-21041 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Reporter: godfrey he >Assignee: godfrey he >Priority: Major > Fix For: 1.13.0 > > > Currently, we use {{List}} to represent the {{ExecNode}} > topology, as we will introduce more features (such as serialize/deserialize > {{ExecNode}}s), It's better we can introduce an unified class to represent > the topology. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Created] (FLINK-21041) Introduce ExecNodeGraph to wrap the ExecNode topology
godfrey he created FLINK-21041: -- Summary: Introduce ExecNodeGraph to wrap the ExecNode topology Key: FLINK-21041 URL: https://issues.apache.org/jira/browse/FLINK-21041 Project: Flink Issue Type: Sub-task Components: Table SQL / Planner Reporter: godfrey he Fix For: 1.13.0 Currently, we use {{List}} to represent the {{ExecNode}} topology, as we will introduce more features (such as serialize/deserialize {{ExecNode}}s), It's better we can introduce an unified class to represent the topology. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-21008) ClusterEntrypoint#shutDownAsync may not be fully executed
[ https://issues.apache.org/jira/browse/FLINK-21008?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268349#comment-17268349 ] Yang Wang commented on FLINK-21008: --- Triggering {{ClusterEntrypoint.closeAsync()}} for the SIGTERM is a better solution. Since ignoring the SIGTERM could cause other issues. For example, deleting a pod gracefully is impossible now. > ClusterEntrypoint#shutDownAsync may not be fully executed > - > > Key: FLINK-21008 > URL: https://issues.apache.org/jira/browse/FLINK-21008 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.11.3, 1.12.1 >Reporter: Yang Wang >Priority: Critical > Fix For: 1.13.0 > > > Recently, in our internal use case for native K8s integration with K8s HA > enabled, we found that the leader related ConfigMaps could be residual in > some corner situations. > After some investigations, I think it is possibly caused by the inappropriate > shutdown process. > In {{ClusterEntrypoint#shutDownAsync}}, we first call the > {{closeClusterComponent}}, which also includes deregistering the Flink > application from cluster management(e.g. Yarn, K8s). Then we call the > {{stopClusterServices}} and {{cleanupDirectories}}. Imagine that the cluster > management do the deregister very fast, the JobManager process receives > SIGNAL 15 before or is being executing the {{stopClusterServices}} and > {{cleanupDirectories}}. The jvm process will directly exit then. So the two > methods may not be executed. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] xintongsong commented on pull request #14697: [FLINK-21024][dist] Keep dynamic properties before job arguments.
xintongsong commented on pull request #14697: URL: https://github.com/apache/flink/pull/14697#issuecomment-763302568 @flinkbot run azure This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-21036) Consider removing automatic configuration fo number of slots from docker
[ https://issues.apache.org/jira/browse/FLINK-21036?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268348#comment-17268348 ] Yang Wang commented on FLINK-21036: --- I am in favor of removing the automatic configuration for {{taskmanager.numberOfTaskSlots}}. Because when we set the cpu limit for docker, we will get an incorrect number. {code:java} $sudo docker run --cpus 1 -it apache/flink:1.12.1 bash -c 'echo $(grep -c ^processor /proc/cpuinfo)' 96 {code} Also for standalone Flink on Kubernetes use case, almost all the users are mounting a ConfigMap[1] for the {{flink-conf.yaml}}. And the {{docker-entrypoint.sh}} do not have the permission to override the conf based the processor. The mounted {{flink-conf.yaml}} from ConfigMap is read-only. I believe we do not expect it to be overridden as well. [1]. https://ci.apache.org/projects/flink/flink-docs-master/deployment/resource-providers/standalone/kubernetes.html#common-cluster-resource-definitions > Consider removing automatic configuration fo number of slots from docker > > > Key: FLINK-21036 > URL: https://issues.apache.org/jira/browse/FLINK-21036 > Project: Flink > Issue Type: Improvement > Components: Deployment / Scripts >Reporter: Chesnay Schepler >Priority: Major > Fix For: 1.13.0 > > > The {{docker-entrypoint.sh}} supports setting the number of task slots via > the {{TASK_MANAGER_NUMBER_OF_TASK_SLOTS}} environment variable, which > defaults to the number of cpu cores via {{$(grep -c ^processor > /proc/cpuinfo)}}. > The environment variable itself is redundant nowadays since we introduced > {{FLINK_PROPERTIES}}, and is no longer documented. > Defaulting to the number of CPU cores can be considered convenience, but it > seems odd to have this specific to docker while the distribution defaults to > {{1}}. > The bigger issue in my mind though is that this creates a configuration > mismatch between the Job- and TaskManager processes; the ResourceManager > specifically needs to know how many slots a worker has to make decisions > about redundancy and allocating resources. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Comment Edited] (FLINK-21032) JsonFileCompactionITCase fails on azure
[ https://issues.apache.org/jira/browse/FLINK-21032?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268345#comment-17268345 ] Leonard Xu edited comment on FLINK-21032 at 1/20/21, 3:19 AM: -- >From the detail log, the job has 3 compact-operator(streaming-writer) to do >the compaction work, but one of them get a checkpoint Exception and did not >finish the checkpoint and did not produce the compaction files as expected. {code:java} 11:36:47,761 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (2/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 1 checkpointing for checkpoint with id=4 (max part counter=66).11:36:47,761 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (3/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 2 checkpointing for checkpoint with id=4 (max part counter=66).11:36:47,762 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (1/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 0 checkpointing for checkpoint with id=4 (max part counter=68).11:36:47,970 [compact-operator (1/3)#0] INFO org.apache.flink.streaming.api.operators.AbstractStreamOperator [] - Compaction time cost is '0.202S', target file is 'file:/tmp/junit8284797133733926241/junit3368655018742742833/compacted-part-10e63de4-6d59-4bfa-9292-c7741c360c96-0-34', input files are '[file:/tmp/junit8284797133733926241/junit3368655018742742833/.uncompacted-part-10e63de4-6d59-4bfa-9292-c7741c360c96-0-34...] '11:36:47,977 [jobmanager-future-thread-26] INFO org.apache.flink.runtime.checkpoint.CheckpointCoordinator[] - Completed checkpoint 4 for job 7c53c3e6a6dd259c6cfa396242680c31 (17514 bytes in 218 ms). 11:36:47,982 [AsyncOperations-thread-1] INFO org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (3/3)#0 - asynchronous part of checkpoint 5 could not be completed.java.util.concurrent.CancellationException: nullat java.util.concurrent.FutureTask.report(FutureTask.java:121) ~[?:1.8.0_275] at java.util.concurrent.FutureTask.get(FutureTask.java:192) ~[?:1.8.0_275] at org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:621) ~[flink-runtime_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]at org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:60) ~[flink-streaming-java_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]at org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:122) [flink-streaming-java_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_275]at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_275]at java.lang.Thread.run(Thread.java:748) [?:1.8.0_275]{code} And thus when read out from the compaction files, we found the data lost. I think this test failed because the source operator(ParallelFiniteTestSource) is still 3 parallelism rather than single parallelism. If the source is multiple parallelism and the source will end in test(in real case, the source won't stop), after one source subtask received endInput,the other source subtask is still waiting checkpoint or running, this situation may lead this problem. How do you think? [~dwysakowicz] [~lzljs3620320] was (Author: leonard xu): >From the detail log, the job has 3 compact-operator(streaming-writer) do the >compaction work, but one of them get a checkpoint Exception and did not finish >the checkpoint and did not produce the compaction files as expected. {code:java} 11:36:47,761 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (2/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 1 checkpointing for checkpoint with id=4 (max part counter=66).11:36:47,761 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (3/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 2 checkpointing for checkpoint with id=4 (max part counter=66).11:36:47,762 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter ->
[jira] [Commented] (FLINK-20500) UpsertKafkaTableITCase.testTemporalJoin test failed
[ https://issues.apache.org/jira/browse/FLINK-20500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268347#comment-17268347 ] Jark Wu commented on FLINK-20500: - Fixed in - master: 59905140b7d6a31f8c07fc8151f33f978e475bae - release-1.12: > UpsertKafkaTableITCase.testTemporalJoin test failed > --- > > Key: FLINK-20500 > URL: https://issues.apache.org/jira/browse/FLINK-20500 > Project: Flink > Issue Type: Bug > Components: Connectors / Kafka, Table SQL / Ecosystem >Affects Versions: 1.12.0, 1.13.0 >Reporter: Huang Xingbo >Assignee: Shengkai Fang >Priority: Major > Labels: pull-request-available, test-stability > Fix For: 1.13.0, 1.12.2 > > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10555=logs=4be4ed2b-549a-533d-aa33-09e28e360cc8=0db94045-2aa0-53fa-f444-0130d6933518] > > {code:java} > 2020-12-06T23:09:20.3744822Z [ERROR] > UpsertKafkaTableITCase.testTemporalJoin:111->temporalJoinUpsertKafka:545 > 2020-12-06T23:09:20.3746021Z Expected: > <10001,100,2020-08-15T00:00:02,Bob,BEIJING,2020-08-15T00:00:01> > 2020-12-06T23:09:20.3747548Z > <10002,101,2020-08-15T00:00:03,Alice,SHANGHAI,2020-08-15T00:00:02> > 2020-12-06T23:09:20.3748539Z > <10002,102,2020-08-15T00:00:04,Greg,BERLIN,2020-08-15T00:00:03> > 2020-12-06T23:09:20.3749294Z > <10002,101,2020-08-16T00:02:01,Alice,WUHAN,2020-08-16T00:02> > 2020-12-06T23:09:20.3749799Z <10004,104,2020-08-16T00:04,null,null,null> > 2020-12-06T23:09:20.3789757Z > <10003,101,2020-08-16T00:04:06,Alice,WUHAN,2020-08-16T00:02> > 2020-12-06T23:09:20.3790891Z > <10004,104,2020-08-16T00:05:06,Tomato,HONGKONG,2020-08-16T00:05:05> > 2020-12-06T23:09:20.3791758Z but: was > <10002,102,2020-08-15T00:00:04,null,null,null> > 2020-12-06T23:09:20.3792494Z <10004,104,2020-08-16T00:04,null,null,null> > 2020-12-06T23:09:20.3793282Z > <10001,100,2020-08-15T00:00:02,Bob,BEIJING,2020-08-15T00:00:01> > 2020-12-06T23:09:20.3794134Z > <10002,101,2020-08-15T00:00:03,Alice,SHANGHAI,2020-08-15T00:00:02> > 2020-12-06T23:09:20.3794928Z > <10002,101,2020-08-16T00:02:01,Alice,WUHAN,2020-08-16T00:02> > 2020-12-06T23:09:20.3795512Z > <10003,101,2020-08-16T00:04:06,Alice,WUHAN,2020-08-16T00:02> > 2020-12-06T23:09:20.3796059Z > <10004,104,2020-08-16T00:05:06,Tomato,HONGKONG,2020-08-16T00:05:05> > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-21039) Broken links in "dev/table/legacy_planner.zh.md"
[ https://issues.apache.org/jira/browse/FLINK-21039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268346#comment-17268346 ] WeiNan Zhao commented on FLINK-21039: - ok,I will solve this problem as soon as possible. > Broken links in "dev/table/legacy_planner.zh.md" > > > Key: FLINK-21039 > URL: https://issues.apache.org/jira/browse/FLINK-21039 > Project: Flink > Issue Type: Bug > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: WeiNan Zhao >Priority: Critical > > dev/table/legacy_planner.zh.md contains some English links causing errors. > {code:java} > Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. > Make sure the document exists and the path is correct. in > dev/table/legacy_planner.zh.md > Could not find document 'dev/batch/index.md' in tag 'link'. > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (FLINK-21032) JsonFileCompactionITCase fails on azure
[ https://issues.apache.org/jira/browse/FLINK-21032?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268345#comment-17268345 ] Leonard Xu commented on FLINK-21032: >From the detail log, the job has 3 compact-operator(streaming-writer) do the >compaction work, but one of them get a checkpoint Exception and did not finish >the checkpoint and did not produce the compaction files as expected. {code:java} 11:36:47,761 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (2/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 1 checkpointing for checkpoint with id=4 (max part counter=66).11:36:47,761 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (3/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 2 checkpointing for checkpoint with id=4 (max part counter=66).11:36:47,762 [Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (1/3)#0] INFO org.apache.flink.streaming.api.functions.sink.filesystem.Buckets [] - Subtask 0 checkpointing for checkpoint with id=4 (max part counter=68).11:36:47,970 [compact-operator (1/3)#0] INFO org.apache.flink.streaming.api.operators.AbstractStreamOperator [] - Compaction time cost is '0.202S', target file is 'file:/tmp/junit8284797133733926241/junit3368655018742742833/compacted-part-10e63de4-6d59-4bfa-9292-c7741c360c96-0-34', input files are '[file:/tmp/junit8284797133733926241/junit3368655018742742833/.uncompacted-part-10e63de4-6d59-4bfa-9292-c7741c360c96-0-34...] '11:36:47,977 [jobmanager-future-thread-26] INFO org.apache.flink.runtime.checkpoint.CheckpointCoordinator[] - Completed checkpoint 4 for job 7c53c3e6a6dd259c6cfa396242680c31 (17514 bytes in 218 ms). 11:36:47,982 [AsyncOperations-thread-1] INFO org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - Source: Custom Source -> SourceConversion(table=[default_catalog.default_database.my_table], fields=[a, b, c]) -> Filter -> streaming-writer (3/3)#0 - asynchronous part of checkpoint 5 could not be completed.java.util.concurrent.CancellationException: nullat java.util.concurrent.FutureTask.report(FutureTask.java:121) ~[?:1.8.0_275] at java.util.concurrent.FutureTask.get(FutureTask.java:192) ~[?:1.8.0_275] at org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:621) ~[flink-runtime_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]at org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:60) ~[flink-streaming-java_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]at org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:122) [flink-streaming-java_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_275]at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_275]at java.lang.Thread.run(Thread.java:748) [?:1.8.0_275]{code} And thus when read out from the compaction files, we found the data lost. I think this test failed because the source operator(ParallelFiniteTestSource) is still 3 parallelism rather than single parallelism. If the source is multiple parallelism and the source will end in test(in real case, the source won't stop), after one source subtask received endInput,the other source subtask is still waiting checkpoint or running, this situation may lead this problem. > JsonFileCompactionITCase fails on azure > --- > > Key: FLINK-21032 > URL: https://issues.apache.org/jira/browse/FLINK-21032 > Project: Flink > Issue Type: Bug > Components: Table SQL / Ecosystem >Reporter: Dawid Wysakowicz >Priority: Blocker > Labels: test-stability > Fix For: 1.13.0 > > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12230=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20 > {code} > [ERROR] Tests run: 3, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 11.39 > s <<< FAILURE! - in org.apache.flink.formats.json.JsonFileCompactionITCase > [ERROR] > testSingleParallelism(org.apache.flink.formats.json.JsonFileCompactionITCase) > Time elapsed: 1.21 s <<< FAILURE! > java.lang.AssertionError: expected:<[+I[0, 0, 0], +I[0, 0, 0], +I[1, 1, 1], > +I[1, 1, 1], +I[2, 2, 2], +I[2, 2, 2], +I[3, 3, 3], +I[3, 3, 3], +I[4, 4, 4], > +I[4, 4, 4], +I[5, 5, 5], +I[5, 5, 5], +I[6, 6, 6], +I[6, 6, 6], +I[7, 7, 7], > +I[7, 7, 7], +I[8,
[GitHub] [flink] wuchong merged pull request #14689: [FLINK-20500][upsert-kafka] Fix temporal join test
wuchong merged pull request #14689: URL: https://github.com/apache/flink/pull/14689 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #14675: [FLINK-18383][docs]translate "JDBC SQL Connector" documentation into …
flinkbot edited a comment on pull request #14675: URL: https://github.com/apache/flink/pull/14675#issuecomment-761769906 ## CI report: * b57985e871e0b703fcaaeb85a597a8ab1584c10b Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12214) * 7b8045005155c54e76d0e19687a1fe9309c04892 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-21039) Broken links in "dev/table/legacy_planner.zh.md"
[ https://issues.apache.org/jira/browse/FLINK-21039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu reassigned FLINK-21039: --- Assignee: WeiNan Zhao > Broken links in "dev/table/legacy_planner.zh.md" > > > Key: FLINK-21039 > URL: https://issues.apache.org/jira/browse/FLINK-21039 > Project: Flink > Issue Type: Bug > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Assignee: WeiNan Zhao >Priority: Critical > > dev/table/legacy_planner.zh.md contains some English links causing errors. > {code:java} > Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. > Make sure the document exists and the path is correct. in > dev/table/legacy_planner.zh.md > Could not find document 'dev/batch/index.md' in tag 'link'. > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] zhuzhurk commented on a change in pull request #14659: [FLINK-20931][coordination] Remove globalModVersion from ExecutionGraph
zhuzhurk commented on a change in pull request #14659: URL: https://github.com/apache/flink/pull/14659#discussion_r560646431 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/executiongraph/ExecutionJobVertex.java ## @@ -499,31 +497,6 @@ public void fail(Throwable t) { } } -public void resetForNewExecution(final long timestamp, final long expectedGlobalModVersion) Review comment: I think `ExecutionVertex#resetForNewExecution(final long timestamp, final long originatingGlobalModVersion)` was used by this method. We should remove it as well. Besides that, could we have a separate hotfix commit to remove this method because it is not the work of removing globalModVersion? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-19446) canal-json has a situation that -U and +U are equal, when updating the null field to be non-null
[ https://issues.apache.org/jira/browse/FLINK-19446?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268342#comment-17268342 ] shizhengchao commented on FLINK-19446: -- [~Balro],i think your patch would be work well. > canal-json has a situation that -U and +U are equal, when updating the null > field to be non-null > > > Key: FLINK-19446 > URL: https://issues.apache.org/jira/browse/FLINK-19446 > Project: Flink > Issue Type: Bug > Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile) >Affects Versions: 1.11.1 >Reporter: shizhengchao >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.13.0 > > > line 118 in CanalJsonDeserializationSchema#deserialize method: > {code:java} > GenericRowData after = (GenericRowData) data.getRow(i, fieldCount); > GenericRowData before = (GenericRowData) old.getRow(i, fieldCount); > for (int f = 0; f < fieldCount; f++) { > if (before.isNullAt(f)) { > // not null fields in "old" (before) means the fields are > changed > // null/empty fields in "old" (before) means the fields are not > changed > // so we just copy the not changed fields into before > before.setField(f, after.getField(f)); > } > } > before.setRowKind(RowKind.UPDATE_BEFORE); > after.setRowKind(RowKind.UPDATE_AFTER); > {code} > if a field is null before update,it will cause -U and +U to be equal -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink-docker] wangyang0918 commented on a change in pull request #59: [FLINK-21037] Deduplicate configuration logic
wangyang0918 commented on a change in pull request #59: URL: https://github.com/apache/flink-docker/pull/59#discussion_r560645542 ## File path: docker-entrypoint.sh ## @@ -112,41 +111,26 @@ if [ "$1" = "help" ]; then elif [ "$1" = "jobmanager" ]; then args=("${args[@]:1}") -prepare_job_manager_start +echo "Starting Job Manager" exec $(drop_privs_cmd) "$FLINK_HOME/bin/jobmanager.sh" start-foreground "${args[@]}" elif [ "$1" = ${COMMAND_STANDALONE} ]; then args=("${args[@]:1}") -prepare_job_manager_start +echo "Starting Job Manager" exec $(drop_privs_cmd) "$FLINK_HOME/bin/standalone-job.sh" start-foreground "${args[@]}" elif [ "$1" = ${COMMAND_HISTORY_SERVER} ]; then args=("${args[@]:1}") echo "Starting History Server" -if [ -n "${FLINK_PROPERTIES}" ]; then -echo "${FLINK_PROPERTIES}" >> "${CONF_FILE}" -fi -envsubst < "${CONF_FILE}" > "${CONF_FILE}.tmp" && mv "${CONF_FILE}.tmp" "${CONF_FILE}" Review comment: So we do not need to do the substitution after then. Why? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] JavaPentesters commented on pull request #14675: [FLINK-18383][docs]translate "JDBC SQL Connector" documentation into …
JavaPentesters commented on pull request #14675: URL: https://github.com/apache/flink/pull/14675#issuecomment-763295166 @leonardBang Thanks for your review! All the mentioned comments have been solved! And, The documentation can be translated to HTML! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-21039) Broken links in "dev/table/legacy_planner.zh.md"
[ https://issues.apache.org/jira/browse/FLINK-21039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268336#comment-17268336 ] Guowei Ma commented on FLINK-21039: --- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12254=logs=6dc02e5c-5865-5c6a-c6c5-92d598e3fc43=ddd6d61a-af16-5d03-2b9a-76a279badf98 > Broken links in "dev/table/legacy_planner.zh.md" > > > Key: FLINK-21039 > URL: https://issues.apache.org/jira/browse/FLINK-21039 > Project: Flink > Issue Type: Bug > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > > dev/table/legacy_planner.zh.md contains some English links causing errors. > {code:java} > Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. > Make sure the document exists and the path is correct. in > dev/table/legacy_planner.zh.md > Could not find document 'dev/batch/index.md' in tag 'link'. > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot edited a comment on pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
flinkbot edited a comment on pull request #14699: URL: https://github.com/apache/flink/pull/14699#issuecomment-763287840 ## CI report: * ddd72a9067d9c7b0fed65b26947d00f530d53310 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=12258) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Reopened] (FLINK-19445) Several tests for HBase connector 1.4 failed with "NoSuchMethodError: com.google.common.base.Preconditions.checkArgument(ZLjava/lang/String;Ljava/lang/Object;)V"
[ https://issues.apache.org/jira/browse/FLINK-19445?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Guowei Ma reopened FLINK-19445: --- These tests fails again in the master. https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12254=logs=ba53eb01-1462-56a3-8e98-0dd97fbcaab5=bfbc6239-57a0-5db0-63f3-41551b4f7d51 > Several tests for HBase connector 1.4 failed with "NoSuchMethodError: > com.google.common.base.Preconditions.checkArgument(ZLjava/lang/String;Ljava/lang/Object;)V" > - > > Key: FLINK-19445 > URL: https://issues.apache.org/jira/browse/FLINK-19445 > Project: Flink > Issue Type: Bug > Components: Connectors / HBase >Affects Versions: 1.12.0 >Reporter: Dian Fu >Assignee: Gyula Fora >Priority: Blocker > Labels: pull-request-available, test-stability > Fix For: 1.12.0 > > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7042=logs=ba53eb01-1462-56a3-8e98-0dd97fbcaab5=bfbc6239-57a0-5db0-63f3-41551b4f7d51 > {code} > 2020-09-28T21:28:29.4171075Z Running > org.apache.flink.connector.hbase1.HBaseTablePlanTest > 2020-09-28T21:28:31.0367584Z Tests run: 5, Failures: 0, Errors: 1, Skipped: > 0, Time elapsed: 1.62 sec <<< FAILURE! - in > org.apache.flink.connector.hbase1.HBaseTablePlanTest > 2020-09-28T21:28:31.0368925Z > testProjectionPushDown(org.apache.flink.connector.hbase1.HBaseTablePlanTest) > Time elapsed: 0.031 sec <<< ERROR! > 2020-09-28T21:28:31.0369805Z org.apache.flink.table.api.ValidationException: > 2020-09-28T21:28:31.0370409Z Unable to create a source for reading table > 'default_catalog.default_database.hTable'. > 2020-09-28T21:28:31.0370707Z > 2020-09-28T21:28:31.0370976Z Table options are: > 2020-09-28T21:28:31.0371204Z > 2020-09-28T21:28:31.0371528Z 'connector'='hbase-1.4' > 2020-09-28T21:28:31.0371871Z 'table-name'='my_table' > 2020-09-28T21:28:31.0372255Z 'zookeeper.quorum'='localhost:2021' > 2020-09-28T21:28:31.0372812Z at > org.apache.flink.table.factories.FactoryUtil.createTableSource(FactoryUtil.java:125) > 2020-09-28T21:28:31.0373359Z at > org.apache.flink.table.planner.plan.schema.CatalogSourceTable.buildTableScan(CatalogSourceTable.scala:135) > 2020-09-28T21:28:31.0373905Z at > org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.scala:78) > 2020-09-28T21:28:31.0374390Z at > org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3492) > 2020-09-28T21:28:31.0375224Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2415) > 2020-09-28T21:28:31.0375867Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2102) > 2020-09-28T21:28:31.0376479Z at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl$$anon$1.convertFrom(FlinkPlannerImpl.scala:181) > 2020-09-28T21:28:31.0377077Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2051) > 2020-09-28T21:28:31.0377593Z at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl$$anon$1.convertFrom(FlinkPlannerImpl.scala:181) > 2020-09-28T21:28:31.0378114Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:661) > 2020-09-28T21:28:31.0378622Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:642) > 2020-09-28T21:28:31.0379132Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3345) > 2020-09-28T21:28:31.0379872Z at > org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:568) > 2020-09-28T21:28:31.0380477Z at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:196) > 2020-09-28T21:28:31.0381128Z at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:154) > 2020-09-28T21:28:31.0381666Z at > org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:823) > 2020-09-28T21:28:31.0382264Z at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:795) > 2020-09-28T21:28:31.0382968Z at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:250) > 2020-09-28T21:28:31.0383550Z at > org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:78) > 2020-09-28T21:28:31.0384172Z at > org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlQuery(TableEnvironmentImpl.java:640) > 2020-09-28T21:28:31.0384700Z at >
[GitHub] [flink-docker] wangyang0918 commented on a change in pull request #57: [FLINK-21034] Refactor jemalloc switch to environment variable
wangyang0918 commented on a change in pull request #57: URL: https://github.com/apache/flink-docker/pull/57#discussion_r560635988 ## File path: docker-entrypoint.sh ## @@ -95,40 +94,34 @@ prepare_job_manager_start() { envsubst < "${CONF_FILE}" > "${CONF_FILE}.tmp" && mv "${CONF_FILE}.tmp" "${CONF_FILE}" } -disable_jemalloc_env() { - # use nameref '_args' to update the passed 'args' within function - local -n _args=$1 - if [ "${_args[0]}" = ${COMMAND_DISABLE_JEMALLOC} ]; then - echo "Disable Jemalloc as the memory allocator" - _args=("${_args[@]:1}") - else - export LD_PRELOAD=$LD_PRELOAD:/usr/lib/x86_64-linux-gnu/libjemalloc.so - fi +maybe_enable_jemalloc() { Review comment: Since the `jemalloc` is enabled by default, maybe the name `disable_jemalloc_if_required` is better. It is aligned with `copy_plugins_if_required`. ## File path: docker-entrypoint.sh ## @@ -95,40 +94,34 @@ prepare_job_manager_start() { envsubst < "${CONF_FILE}" > "${CONF_FILE}.tmp" && mv "${CONF_FILE}.tmp" "${CONF_FILE}" } -disable_jemalloc_env() { - # use nameref '_args' to update the passed 'args' within function - local -n _args=$1 - if [ "${_args[0]}" = ${COMMAND_DISABLE_JEMALLOC} ]; then - echo "Disable Jemalloc as the memory allocator" - _args=("${_args[@]:1}") - else - export LD_PRELOAD=$LD_PRELOAD:/usr/lib/x86_64-linux-gnu/libjemalloc.so - fi +maybe_enable_jemalloc() { +if ! [ "${DISABLE_JEMALLOC:-false}" == "true" ]; then Review comment: ```suggestion if [ "${DISABLE_JEMALLOC:-false}" != "true" ]; then ``` Or ```suggestion if [ "${DISABLE_JEMALLOC:-false}" == "false" ]; then ``` ## File path: testing/bin/docker-entrypoint.sh ## @@ -0,0 +1,37 @@ +#!/bin/bash -e + +### +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +### + +if ! [ "$1" == "hello" ] || ! [ "$2" == "world" ]; then +echo "Incorrect arguments passed through:" "$@" +exit 1 +fi + +originalLdPreloadSetting=$3 +jemallocDisabled=$4 + +if [ "$jemallocDisabled" == "true" ] && ! [ "$originalLdPreloadSetting" == "$LD_PRELOAD" ]; then +echo "jemalloc was not disabled; exploed LD_PRELOAD to be '$originalLdPreloadSetting' but was '$LD_PRELOAD'" Review comment: typo here? `exploed` -> `expected` This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] JavaPentesters commented on a change in pull request #14675: [FLINK-18383][docs]translate "JDBC SQL Connector" documentation into …
JavaPentesters commented on a change in pull request #14675: URL: https://github.com/apache/flink/pull/14675#discussion_r560637320 ## File path: docs/dev/table/connectors/jdbc.zh.md ## @@ -31,35 +31,35 @@ under the License. * This will be replaced by the TOC {:toc} -The JDBC connector allows for reading data from and writing data into any relational databases with a JDBC driver. This document describes how to setup the JDBC connector to run SQL queries against relational databases. +JDBC连接器允许使用JDBC驱动向任意类型的关系型数据库读取或者写入数据。本文档描述了针对关系型数据库如何通过建立JDBC连接器来执行SQL查询。 -The JDBC sink operate in upsert mode for exchange UPDATE/DELETE messages with the external system if a primary key is defined on the DDL, otherwise, it operates in append mode and doesn't support to consume UPDATE/DELETE messages. +如果一个主键定义在 DDL 中,JDBC sink 将以 upsert 模式与外部系统交换 UPDATE/DELETE 消息;否则,它将以appened 模式与外部系统交换消息且不支持消费 UPDATE/DELETE 消息。 -Dependencies + +依赖 {% assign connector = site.data.sql-connectors['jdbc'] %} {% include sql-connector-download-table.zh.html -connector=connector +connector=connector Review comment: Thanks! Got it! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-20254) HiveTableSourceITCase.testStreamPartitionReadByCreateTime times out
[ https://issues.apache.org/jira/browse/FLINK-20254?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17268331#comment-17268331 ] Guowei Ma commented on FLINK-20254: --- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=12254=logs=245e1f2e-ba5b-5570-d689-25ae21e5302f=e7f339b2-a7c3-57d9-00af-3712d4b15354 > HiveTableSourceITCase.testStreamPartitionReadByCreateTime times out > --- > > Key: FLINK-20254 > URL: https://issues.apache.org/jira/browse/FLINK-20254 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive >Affects Versions: 1.12.0 >Reporter: Robert Metzger >Assignee: Leonard Xu >Priority: Critical > Labels: test-stability > Fix For: 1.13.0 > > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=9808=logs=fc5181b0-e452-5c8f-68de-1097947f6483=62110053-334f-5295-a0ab-80dd7e2babbf > {code} > 2020-11-19T10:34:23.5591765Z [ERROR] Tests run: 18, Failures: 0, Errors: 1, > Skipped: 0, Time elapsed: 192.243 s <<< FAILURE! - in > org.apache.flink.connectors.hive.HiveTableSourceITCase > 2020-11-19T10:34:23.5593193Z [ERROR] > testStreamPartitionReadByCreateTime(org.apache.flink.connectors.hive.HiveTableSourceITCase) > Time elapsed: 120.075 s <<< ERROR! > 2020-11-19T10:34:23.5593929Z org.junit.runners.model.TestTimedOutException: > test timed out after 12 milliseconds > 2020-11-19T10:34:23.5594321Z at java.lang.Thread.sleep(Native Method) > 2020-11-19T10:34:23.5594777Z at > org.apache.flink.streaming.api.operators.collect.CollectResultFetcher.sleepBeforeRetry(CollectResultFetcher.java:231) > 2020-11-19T10:34:23.5595378Z at > org.apache.flink.streaming.api.operators.collect.CollectResultFetcher.next(CollectResultFetcher.java:119) > 2020-11-19T10:34:23.5596001Z at > org.apache.flink.streaming.api.operators.collect.CollectResultIterator.nextResultFromFetcher(CollectResultIterator.java:103) > 2020-11-19T10:34:23.5596610Z at > org.apache.flink.streaming.api.operators.collect.CollectResultIterator.hasNext(CollectResultIterator.java:77) > 2020-11-19T10:34:23.5597218Z at > org.apache.flink.table.planner.sinks.SelectTableSinkBase$RowIteratorWrapper.hasNext(SelectTableSinkBase.java:115) > 2020-11-19T10:34:23.5597811Z at > org.apache.flink.table.api.internal.TableResultImpl$CloseableRowIteratorWrapper.hasNext(TableResultImpl.java:355) > 2020-11-19T10:34:23.5598555Z at > org.apache.flink.connectors.hive.HiveTableSourceITCase.fetchRows(HiveTableSourceITCase.java:653) > 2020-11-19T10:34:23.5599407Z at > org.apache.flink.connectors.hive.HiveTableSourceITCase.testStreamPartitionReadByCreateTime(HiveTableSourceITCase.java:594) > 2020-11-19T10:34:23.5599982Z at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2020-11-19T10:34:23.5600393Z at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2020-11-19T10:34:23.5600865Z at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2020-11-19T10:34:23.5601300Z at > java.lang.reflect.Method.invoke(Method.java:498) > 2020-11-19T10:34:23.5601713Z at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) > 2020-11-19T10:34:23.5602211Z at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2020-11-19T10:34:23.5602688Z at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) > 2020-11-19T10:34:23.5603181Z at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2020-11-19T10:34:23.5603753Z at > org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298) > 2020-11-19T10:34:23.5604308Z at > org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292) > 2020-11-19T10:34:23.5604780Z at > java.util.concurrent.FutureTask.run(FutureTask.java:266) > 2020-11-19T10:34:23.5605114Z at java.lang.Thread.run(Thread.java:748) > 2020-11-19T10:34:23.5605299Z > 2020-11-19T10:34:24.4180149Z [INFO] Running > org.apache.flink.connectors.hive.TableEnvHiveConnectorITCase > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] flinkbot commented on pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
flinkbot commented on pull request #14699: URL: https://github.com/apache/flink/pull/14699#issuecomment-763287840 ## CI report: * ddd72a9067d9c7b0fed65b26947d00f530d53310 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run travis` re-run the last Travis build - `@flinkbot run azure` re-run the last Azure build This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
flinkbot commented on pull request #14699: URL: https://github.com/apache/flink/pull/14699#issuecomment-763280085 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit ddd72a9067d9c7b0fed65b26947d00f530d53310 (Wed Jan 20 02:15:32 UTC 2021) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-21011) Separate the implementation of StreamExecIntervalJoin
[ https://issues.apache.org/jira/browse/FLINK-21011?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-21011: --- Labels: pull-request-available (was: ) > Separate the implementation of StreamExecIntervalJoin > - > > Key: FLINK-21011 > URL: https://issues.apache.org/jira/browse/FLINK-21011 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Reporter: Wenlong Lyu >Assignee: Wenlong Lyu >Priority: Major > Labels: pull-request-available > Fix For: 1.13.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[GitHub] [flink] wenlong88 opened a new pull request #14699: [FLINK-21011][table-planner-blink] Separate implementation of StreamExecIntervalJoin
wenlong88 opened a new pull request #14699: URL: https://github.com/apache/flink/pull/14699 ## What is the purpose of the change Introduce StreamPhysicalIntervalJoin and make StreamExecIntervalJoin only extends from ExecNode ## Brief change log 1. introduce StreamPhysicalIntervalJoin and make StreamExecIntervalJoin only extends from ExecNode 2. introduce IntervalJoinFunction to wrapper non-equil JoinCondition and null join key filter in join spec, so that we do not need to compare join keys in runtime because the join keys is already compared by the group key in runtime. ## Verifying this change This change is a refactor work, the change is already covered by existing tests. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-21040) The exception of "Unknown consumerTag" may throw when RMQSource close() is called.
pcalpha created FLINK-21040: --- Summary: The exception of "Unknown consumerTag" may throw when RMQSource close() is called. Key: FLINK-21040 URL: https://issues.apache.org/jira/browse/FLINK-21040 Project: Flink Issue Type: Bug Components: Connectors/ RabbitMQ Affects Versions: 1.12.1, 1.11.3, 1.11.2, 1.11.1, 1.12.0, 1.10.2, 1.11.0, 1.10.1, 1.10.0, 1.9.3, 1.9.2, 1.9.1, 1.9.0, 1.8.3, 1.8.2, 1.8.1, 1.8.0, 1.7.2, 1.6.4, 1.6.3 Reporter: pcalpha The log is below 2020-10-27 21:00:57 [ Source: rabbitmqSource -> Flat Map (1/1):292311427 ] - [ ERROR ] Error during disposal of stream operator.2020-10-27 21:00:57 [ Source: rabbitmqSource -> Flat Map (1/1):292311427 ] - [ ERROR ] Error during disposal of stream operator.java.lang.RuntimeException: Error while cancelling RMQ consumer on xxxQueueName at 10.xx.xx.102 at org.apache.flink.streaming.connectors.rabbitmq.RMQSource.close(RMQSource.java:196) at org.apache.flink.api.common.functions.util.FunctionUtils.closeFunction(FunctionUtils.java:43) at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.dispose(AbstractUdfStreamOperator.java:117) at org.apache.flink.streaming.runtime.tasks.StreamTask.disposeAllOperators(StreamTask.java:703) at org.apache.flink.streaming.runtime.tasks.StreamTask.cleanUpInvoke(StreamTask.java:635) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:542) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:721) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:546) at java.base/java.lang.Thread.run(Thread.java:834)Caused by: java.io.IOException: Unknown consumerTag at com.rabbitmq.client.impl.ChannelN.basicCancel(ChannelN.java:1285) at com.rabbitmq.client.impl.recovery.AutorecoveringChannel.basicCancel(AutorecoveringChannel.java:482) at org.apache.flink.streaming.connectors.rabbitmq.RMQSource.close(RMQSource.java:192) ... 8 more -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Assigned] (FLINK-21011) Separate the implementation of StreamExecIntervalJoin
[ https://issues.apache.org/jira/browse/FLINK-21011?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he reassigned FLINK-21011: -- Assignee: Wenlong Lyu > Separate the implementation of StreamExecIntervalJoin > - > > Key: FLINK-21011 > URL: https://issues.apache.org/jira/browse/FLINK-21011 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Reporter: Wenlong Lyu >Assignee: Wenlong Lyu >Priority: Major > Fix For: 1.13.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Assigned] (FLINK-20968) Remove legacy exec nodes
[ https://issues.apache.org/jira/browse/FLINK-20968?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he reassigned FLINK-20968: -- Assignee: godfrey he > Remove legacy exec nodes > > > Key: FLINK-20968 > URL: https://issues.apache.org/jira/browse/FLINK-20968 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Reporter: godfrey he >Assignee: godfrey he >Priority: Major > Fix For: 1.13.0 > > -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Updated] (FLINK-21039) Broken links in "dev/table/legacy_planner.zh.md"
[ https://issues.apache.org/jira/browse/FLINK-21039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo updated FLINK-21039: - Description: dev/table/legacy_planner.zh.md contains some English links causing errors. {code:java} Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. Make sure the document exists and the path is correct. in dev/table/legacy_planner.zh.md Could not find document 'dev/batch/index.md' in tag 'link'. {code} was: dev/table/legacy_planner.zh.md contains some Chinese links causing errors. {code:java} Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. Make sure the document exists and the path is correct. in dev/table/legacy_planner.zh.md Could not find document 'dev/batch/index.md' in tag 'link'. {code} > Broken links in "dev/table/legacy_planner.zh.md" > > > Key: FLINK-21039 > URL: https://issues.apache.org/jira/browse/FLINK-21039 > Project: Flink > Issue Type: Bug > Components: Documentation, Table SQL / Ecosystem >Affects Versions: 1.13.0 >Reporter: Huang Xingbo >Priority: Critical > > dev/table/legacy_planner.zh.md contains some English links causing errors. > {code:java} > Liquid Exception: Could not find document 'dev/batch/index.md' in tag 'link'. > Make sure the document exists and the path is correct. in > dev/table/legacy_planner.zh.md > Could not find document 'dev/batch/index.md' in tag 'link'. > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)