[GitHub] [flink] flinkbot edited a comment on pull request #18346: [DO NOT MERGE][FLINK-25085] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
flinkbot edited a comment on pull request #18346: URL: https://github.com/apache/flink/pull/18346#issuecomment-1011842175 ## CI report: * 195e75ed899f947fbb8b535f9bfc16f501d0f86e Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29363) * d5e74a33d64d43c2d0b5c45551de399f880602ff Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29366) * b96af7d7646dae5707dbc318239a457d328e926a UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) * 111d05bc4d409c357223c238e67bcb34880a1951 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lichang-bd commented on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
lichang-bd commented on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011883993 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18346: [DO NOT MERGE][FLINK-25085] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
flinkbot edited a comment on pull request #18346: URL: https://github.com/apache/flink/pull/18346#issuecomment-1011842175 ## CI report: * 195e75ed899f947fbb8b535f9bfc16f501d0f86e Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29363) * d5e74a33d64d43c2d0b5c45551de399f880602ff Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29366) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) * 8337b385d6b96699e33fe7668d8b34bfc38beb2d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18102: [FLINK-25033][runtime] Let some scheduler components updatable
flinkbot edited a comment on pull request #18102: URL: https://github.com/apache/flink/pull/18102#issuecomment-993332110 ## CI report: * 114a3c714e91f1075ddc647417ff59bc1d2cc907 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29365) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] fapaul commented on a change in pull request #18302: [FLINK-25569][core] Add decomposed Sink V2 interface
fapaul commented on a change in pull request #18302: URL: https://github.com/apache/flink/pull/18302#discussion_r783706614 ## File path: flink-core/src/main/java/org/apache/flink/api/connector/sink2/StatefulSink.java ## @@ -0,0 +1,97 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.api.connector.sink2; + +import org.apache.flink.annotation.PublicEvolving; +import org.apache.flink.core.io.SimpleVersionedSerializer; + +import java.io.IOException; +import java.util.Collection; +import java.util.List; + +/** + * A {@link Sink} with a stateful {@link SinkWriter}. + * + * The {@link StatefulSink} needs to be serializable. All configuration should be validated + * eagerly. The respective sink writers are transient and will only be created in the subtasks on + * the taskmanagers. + * + * @param The type of the sink's input + * @param The type of the sink writer's state + */ +@PublicEvolving +public interface StatefulSink extends Sink { + +/** + * Create a {@link StatefulSinkWriter}. + * + * @param context the runtime context. + * @return A sink writer. + * @throws IOException for any failure during creation. + */ +StatefulSinkWriter createWriter(InitContext context) throws IOException; + +/** + * Create a {@link StatefulSinkWriter} from a recovered state. + * + * @param context the runtime context. + * @return A sink writer. + * @throws IOException for any failure during creation. + */ +StatefulSinkWriter restoreWriter( +InitContext context, Collection recoveredState) throws IOException; + +/** + * Any stateful sink needs to provide this state serializer and implement {@link + * StatefulSinkWriter#snapshotState(long)} properly. The respective state is used in {@link + * #restoreWriter(InitContext, Collection)} on recovery. + * + * @return the serializer of the writer's state type. + */ +SimpleVersionedSerializer getWriterStateSerializer(); Review comment: Removing all these optionals was one of the intentions behind designing the new interfaces. Sink developers can now explicitly decide which functionality they want to support and implement the interfaces accordingly [1]. With the Sink V1 interfaces they basically always had to implement everything except that some of the methods have default implementations. [1] https://cwiki.apache.org/confluence/display/FLINK/FLIP-191%3A+Extend+unified+Sink+interface+to+support+small+file+compaction#FLIP191:ExtendunifiedSinkinterfacetosupportsmallfilecompaction-SimpleSink -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] fapaul commented on a change in pull request #18302: [FLINK-25569][core] Add decomposed Sink V2 interface
fapaul commented on a change in pull request #18302: URL: https://github.com/apache/flink/pull/18302#discussion_r783704812 ## File path: flink-core/src/main/java/org/apache/flink/api/connector/sink2/Committer.java ## @@ -0,0 +1,103 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.api.connector.sink2; + +import org.apache.flink.annotation.PublicEvolving; + +import java.io.IOException; +import java.util.Collection; + +/** + * The {@code Committer} is responsible for committing the data staged by the {@link + * TwoPhaseCommittingSink.PrecommittingSinkWriter} in the second step of a two-phase commit + * protocol. + * + * A commit must be idempotent: If some failure occurs in Flink during commit phase, Flink will + * restart from previous checkpoint and re-attempt to commit all committables. Thus, some or all + * committables may have already been committed. These {@link CommitRequest}s must not change the + * external system and implementers are asked to signal {@link + * CommitRequest#signalAlreadyCommitted()}. + * + * @param The type of information needed to commit the staged data + */ +@PublicEvolving +public interface Committer extends AutoCloseable { +/** + * Commit the given list of {@link CommT}. + * + * @param committables A list of commit requests staged by the sink writer. + * @throws IOException for reasons that may yield a complete restart of the job. + */ +void commit(Collection> committables) +throws IOException, InterruptedException; + +/** + * A request to commit a specific committable. + * + * @param + */ +@PublicEvolving +interface CommitRequest { + +/** Returns the committable. */ +CommT getCommittable(); + +/** + * Returns how many times this particular committable has been retried. Starts at 0 for the + * first attempt. + */ +int getNumberOfRetries(); + +/** + * The commit failed for known reason and should not be retried. + * + * Currently calling this method only logs the error, discards the comittable and + * continues. In the future the behaviour might be configurable. + */ +void signalFailedWithKnownReason(Throwable t); + +/** + * The commit failed for unknown reason and should not be retried. + * + * Currently calling this method fails the job. In the future the behaviour might be + * configurable. + */ +void signalFailedWithUnknownReason(Throwable t); + +/** + * The commit failed for a retriable reason. If the sink supports a retry maximum, this may + * permanently fail after reaching that maximum. Else the committable will be retried as + * long as this method is invoked after each attempt. + */ +void retryLater(); Review comment: So far none of our sinks sets a number of maximum retries but in the future, we might consider it. The retry mechanism will work internally similar to the current implementation [1]. As soon as the committable is retried we enqueue in the mailbox that is polled "periodically" and retried. Moreover during the next checkpoint, the committable is retried as well. [1] https://github.com/apache/flink/blob/dbbf2a36111da1faea5c901e3b008cc788913bf8/flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/sink/CommitterOperator.java#L96 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18346: [DO NOT MERGE][FLINK-25085] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
flinkbot edited a comment on pull request #18346: URL: https://github.com/apache/flink/pull/18346#issuecomment-1011842175 ## CI report: * 195e75ed899f947fbb8b535f9bfc16f501d0f86e Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29363) * d5e74a33d64d43c2d0b5c45551de399f880602ff UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-25615) FlinkKafkaProducer fail to correctly migrate pre Flink 1.9 state
[ https://issues.apache.org/jira/browse/FLINK-25615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475159#comment-17475159 ] Martijn Visser commented on FLINK-25615: [~Matthias Schwalbe] Definitely. Thanks for reaching out! > FlinkKafkaProducer fail to correctly migrate pre Flink 1.9 state > > > Key: FLINK-25615 > URL: https://issues.apache.org/jira/browse/FLINK-25615 > Project: Flink > Issue Type: Bug > Components: Connectors / Kafka >Affects Versions: 1.9.0 >Reporter: Matthias Schwalbe >Priority: Major > > I've found an unnoticed error in FlinkKafkaProvider when migrating from pre > Flink 1.9 state to versions starting with Flink 1.9: > * the operator state for next-transactional-id-hint should be deleted and > replaced by operator state next-transactional-id-hint-v2, however > * operator state next-transactional-id-hint is never deleted > * see here: [1] : > {quote} if (context.getOperatorStateStore() > .getRegisteredStateNames() > .contains(NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR)) { > migrateNextTransactionalIdHindState(context); > }{quote} * migrateNextTransactionalIdHindState is never called, as > the condition cannot become true: > ** getRegisteredStateNames returns a list of String, whereas > NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR is ListStateDescriptor (type mismatch) > The Effect is: > * because NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR is for a UnionListState, and > * the state is not cleared, > * each time the job restarts from a savepoint or checkpoint the size > multiplies times the parallelism > * then because each entry leaves an offset in metadata, akka.framesize > becomes too small, before we run into memory overflow > > The breaking change has been introduced in commit > 70fa80e3862b367be22b593db685f9898a2838ef > > A simple fix would be to change the code to: > {quote} if (context.getOperatorStateStore() > .getRegisteredStateNames() > .contains(NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR.getName())) { > migrateNextTransactionalIdHindState(context); > } > {quote} > > Although FlinkKafkaProvider is marked as deprecated it is probably a while > here to stay > > Greeting > Matthias (Thias) Schwalbe > > [1] > https://github.com/apache/flink/blob/d7cf2c10f8d4fba81173854cbd8be27c657c7c7f/flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer.java#L1167-L1171 > -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18102: [FLINK-25033][runtime] Let some scheduler components updatable
flinkbot edited a comment on pull request #18102: URL: https://github.com/apache/flink/pull/18102#issuecomment-993332110 ## CI report: * 3bb15a5495d1626bded0929f766c179edfafd995 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=28557) * 114a3c714e91f1075ddc647417ff59bc1d2cc907 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29365) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-25628) Introduce RecordReader and related classes for table store
[ https://issues.apache.org/jira/browse/FLINK-25628?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-25628: --- Labels: pull-request-available (was: ) > Introduce RecordReader and related classes for table store > -- > > Key: FLINK-25628 > URL: https://issues.apache.org/jira/browse/FLINK-25628 > Project: Flink > Issue Type: Sub-task > Components: Table Store >Reporter: Jingsong Lee >Assignee: Jingsong Lee >Priority: Major > Labels: pull-request-available > Fix For: table-store-0.1.0 > > > - Introduce RecordReader interface: The reader that reads the batches of > records. > - Introduce SortMergeReader: This reader is to read a list of `RecordReader`, > which is already sorted by key and sequence number, and perform a sort merge > algorithm. `KeyValue` with the same key will also be combined during sort > merging. > - Introduce ConcatRecordReader: This reader is to concatenate a list of > `RecordReader` and read them sequentially. The input list is already sorted > by key and sequence number, and the key intervals do not overlap each other. > - Introduce FieldStats: Statistics for each field. > - Introduce SstPathFactory: Factory which produces new Path for sst files. > - Introduce SstFile and SstFileMeta: This SstFile includes several > `KeyValue`, representing the changes inserted into the file storage. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Updated] (FLINK-25628) Introduce RecordReader and related classes for table store
[ https://issues.apache.org/jira/browse/FLINK-25628?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee updated FLINK-25628: - Description: - Introduce RecordReader interface: The reader that reads the batches of records. - Introduce SortMergeReader: This reader is to read a list of `RecordReader`, which is already sorted by key and sequence number, and perform a sort merge algorithm. `KeyValue` with the same key will also be combined during sort merging. - Introduce ConcatRecordReader: This reader is to concatenate a list of `RecordReader` and read them sequentially. The input list is already sorted by key and sequence number, and the key intervals do not overlap each other. - Introduce FieldStats: Statistics for each field. - Introduce SstPathFactory: Factory which produces new Path for sst files. - Introduce SstFile and SstFileMeta: This SstFile includes several `KeyValue`, representing the changes inserted into the file storage. > Introduce RecordReader and related classes for table store > -- > > Key: FLINK-25628 > URL: https://issues.apache.org/jira/browse/FLINK-25628 > Project: Flink > Issue Type: Sub-task > Components: Table Store >Reporter: Jingsong Lee >Assignee: Jingsong Lee >Priority: Major > Fix For: table-store-0.1.0 > > > - Introduce RecordReader interface: The reader that reads the batches of > records. > - Introduce SortMergeReader: This reader is to read a list of `RecordReader`, > which is already sorted by key and sequence number, and perform a sort merge > algorithm. `KeyValue` with the same key will also be combined during sort > merging. > - Introduce ConcatRecordReader: This reader is to concatenate a list of > `RecordReader` and read them sequentially. The input list is already sorted > by key and sequence number, and the key intervals do not overlap each other. > - Introduce FieldStats: Statistics for each field. > - Introduce SstPathFactory: Factory which produces new Path for sst files. > - Introduce SstFile and SstFileMeta: This SstFile includes several > `KeyValue`, representing the changes inserted into the file storage. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink-web] gaoyunhaii closed pull request #499: Rebuild website
gaoyunhaii closed pull request #499: URL: https://github.com/apache/flink-web/pull/499 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-table-store] JingsongLi opened a new pull request #4: [FLINK-25628] Introduce RecordReader and related classes for table store
JingsongLi opened a new pull request #4: URL: https://github.com/apache/flink-table-store/pull/4 - Introduce RecordReader interface: The reader that reads the batches of records. - Introduce SortMergeReader: This reader is to read a list of `RecordReader`, which is already sorted by key and sequence number, and perform a sort merge algorithm. `KeyValue` with the same key will also be combined during sort merging. - Introduce ConcatRecordReader: This reader is to concatenate a list of `RecordReader` and read them sequentially. The input list is already sorted by key and sequence number, and the key intervals do not overlap each other. - Introduce FieldStats: Statistics for each field. - Introduce SstPathFactory: Factory which produces new Path for sst files. - Introduce SstFile and SstFileMeta: This SstFile includes several `KeyValue`, representing the changes inserted into the file storage. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-25627) Add basic structures of file store in table-store
[ https://issues.apache.org/jira/browse/FLINK-25627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee closed FLINK-25627. Resolution: Fixed master: 40b8a57e71ca833bb5e7045d06b4621010762435 bda887cec0b78f991663e140b26888c507964664 8845586c37550b6b56446ad8208b732650fe426a > Add basic structures of file store in table-store > - > > Key: FLINK-25627 > URL: https://issues.apache.org/jira/browse/FLINK-25627 > Project: Flink > Issue Type: Sub-task > Components: Table Store >Reporter: Jingsong Lee >Assignee: Jingsong Lee >Priority: Major > Labels: pull-request-available > Fix For: table-store-0.1.0 > > > Add basic structures of file store in table-store: > * Add OffsetRowData > * Add KeyValue and KeyValueSerializer > * Add MemTable and Accumulator -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) * 8337b385d6b96699e33fe7668d8b34bfc38beb2d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18102: [FLINK-25033][runtime] Let some scheduler components updatable
flinkbot edited a comment on pull request #18102: URL: https://github.com/apache/flink/pull/18102#issuecomment-993332110 ## CI report: * 3bb15a5495d1626bded0929f766c179edfafd995 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=28557) * 114a3c714e91f1075ddc647417ff59bc1d2cc907 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-25638) Increase the default write buffer size of sort-shuffle to 16M
[ https://issues.apache.org/jira/browse/FLINK-25638?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao reassigned FLINK-25638: --- Assignee: Yingjie Cao > Increase the default write buffer size of sort-shuffle to 16M > - > > Key: FLINK-25638 > URL: https://issues.apache.org/jira/browse/FLINK-25638 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Network >Reporter: Yingjie Cao >Assignee: Yingjie Cao >Priority: Major > Fix For: 1.15.0 > > > As discussed in > [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this > ticket aims to increase the default write buffer size of sort-shuffle to 16M. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Assigned] (FLINK-25637) Make sort-shuffle the default shuffle implementation for batch jobs
[ https://issues.apache.org/jira/browse/FLINK-25637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao reassigned FLINK-25637: --- Assignee: Yingjie Cao > Make sort-shuffle the default shuffle implementation for batch jobs > --- > > Key: FLINK-25637 > URL: https://issues.apache.org/jira/browse/FLINK-25637 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Network >Reporter: Yingjie Cao >Assignee: Yingjie Cao >Priority: Major > Fix For: 1.15.0 > > > As discussed in > [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this > ticket aims to make sort-shuffle the default shuffle implementation for batch > jobs. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Assigned] (FLINK-25639) Increase the default read buffer size of sort-shuffle to 64M
[ https://issues.apache.org/jira/browse/FLINK-25639?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao reassigned FLINK-25639: --- Assignee: Yingjie Cao > Increase the default read buffer size of sort-shuffle to 64M > > > Key: FLINK-25639 > URL: https://issues.apache.org/jira/browse/FLINK-25639 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Network >Reporter: Yingjie Cao >Assignee: Yingjie Cao >Priority: Major > Fix For: 1.15.0 > > > As discussed in > [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this > ticket aims to increase the default read buffer size of sort-shuffle to 64M. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Updated] (FLINK-25627) Add basic structures of file store in table-store
[ https://issues.apache.org/jira/browse/FLINK-25627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-25627: --- Labels: pull-request-available (was: ) > Add basic structures of file store in table-store > - > > Key: FLINK-25627 > URL: https://issues.apache.org/jira/browse/FLINK-25627 > Project: Flink > Issue Type: Sub-task > Components: Table Store >Reporter: Jingsong Lee >Assignee: Jingsong Lee >Priority: Major > Labels: pull-request-available > Fix For: table-store-0.1.0 > > > Add basic structures of file store in table-store: > * Add OffsetRowData > * Add KeyValue and KeyValueSerializer > * Add MemTable and Accumulator -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink-table-store] JingsongLi merged pull request #3: [FLINK-25627] Add basic structures of file store in table-store
JingsongLi merged pull request #3: URL: https://github.com/apache/flink-table-store/pull/3 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-24207) Add support of KeyedState in TwoPhaseCommitSinkFunction
[ https://issues.apache.org/jira/browse/FLINK-24207?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475155#comment-17475155 ] zlzhang0122 commented on FLINK-24207: - [~roman] ok, I will check for this, thx!! > Add support of KeyedState in TwoPhaseCommitSinkFunction > --- > > Key: FLINK-24207 > URL: https://issues.apache.org/jira/browse/FLINK-24207 > Project: Flink > Issue Type: New Feature > Components: API / DataStream >Affects Versions: 1.12.2, 1.13.1 >Reporter: zlzhang0122 >Priority: Major > > Now, the implementation of TwoPhaseCommitSinkFunction is based on operator > state, but operator state will do a deep copy when taking checkpoint, so > large operator state may produce a OOM error. Add support of KeyedState in > TwoPhaseCommitSinkFunction maybe a good choice to avoid the OOM error and > give users more convenience. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Assigned] (FLINK-25484) TableRollingPolicy do not support inactivityInterval config which is supported in datastream api
[ https://issues.apache.org/jira/browse/FLINK-25484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee reassigned FLINK-25484: Assignee: LiChang > TableRollingPolicy do not support inactivityInterval config which is > supported in datastream api > > > Key: FLINK-25484 > URL: https://issues.apache.org/jira/browse/FLINK-25484 > Project: Flink > Issue Type: Improvement > Components: Connectors / FileSystem, Table SQL / Ecosystem >Affects Versions: 1.15.0 >Reporter: LiChang >Assignee: LiChang >Priority: Major > Fix For: 1.15.0 > > Attachments: image-2022-01-13-11-36-24-102.png > > > TableRollingPolicy do not support inactivityInterval config > public static class TableRollingPolicy extends > CheckpointRollingPolicy { > private final boolean rollOnCheckpoint; > private final long rollingFileSize; > private final long rollingTimeInterval; > public TableRollingPolicy( > boolean rollOnCheckpoint, long rollingFileSize, long rollingTimeInterval) { > this.rollOnCheckpoint = rollOnCheckpoint; > Preconditions.checkArgument(rollingFileSize > 0L); > Preconditions.checkArgument(rollingTimeInterval > 0L); > this.rollingFileSize = rollingFileSize; > this.rollingTimeInterval = rollingTimeInterval; > } > -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) * 8337b385d6b96699e33fe7668d8b34bfc38beb2d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-25057) Streaming File Sink writing to HDFS
[ https://issues.apache.org/jira/browse/FLINK-25057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475147#comment-17475147 ] hanjie commented on FLINK-25057: hi,[~zhisheng], we can set file PartPrefix. {color:#0052cc}for {color}_example: part-uuid-_ {_}when we restart work, t{_}he file name will not be duplicated. > Streaming File Sink writing to HDFS > > > Key: FLINK-25057 > URL: https://issues.apache.org/jira/browse/FLINK-25057 > Project: Flink > Issue Type: Improvement > Components: Connectors / FileSystem >Affects Versions: 1.12.1 >Reporter: hanjie >Priority: Major > > Env: Flink 1.12.1 > kafka --> hdfs > hdfs : Streaming File Sink > When I first start flink task: > *First part file example:* > part-0-0 > part-0-1 > .part-0-2.inprogress.952eb958-dac9-4f2c-b92f-9084ed536a1c > I cancel flink task. then, i restart task without savepoint or checkpoint. > Task run for a while. > *Second part file example:* > part-0-0 > part-0-1 > .part-0-2.inprogress.952eb958-dac9-4f2c-b92f-9084ed536a1c > .part-0-0.inprogress.0e2f234b-042d-4232-a5f7-c980f04ca82d > 'part-0-2.inprogress.952eb958-dac9-4f2c-b92f-9084ed536a1c' not rename > file and bucketIndex will start zero. > I view related code. Start task need savepoint or checkpoint. I choose > savepoint.The above question disappears, when i start third test. > But, if i use expire savepoint. Task will throw exception. > java.io.FileNotFoundException: File does not exist: > /ns-hotel/hotel_sa_log/stream/sa_cpc_ad_log_list_detail_dwd/2021-11-25/.part-6-1537.inprogress.cd9c756a-1756-4dc5-9325-485fe99a2803\n\tat > > org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1309)\n\tat > > org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301)\n\tat > > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)\n\tat > > org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1317)\n\tat > org.apache.hadoop.fs.FileSystem.resolvePath(FileSystem.java:752)\n\tat > org.apache.hadoop.fs.FilterFileSystem.resolvePath(FilterFileSystem.java:153)\n\tat > > org.apache.hadoop.fs.viewfs.ChRootedFileSystem.resolvePath(ChRootedFileSystem.java:373)\n\tat > > org.apache.hadoop.fs.viewfs.ViewFileSystem.resolvePath(ViewFileSystem.java:243)\n\tat > > org.apache.flink.runtime.fs.hdfs.HadoopRecoverableFsDataOutputStream.revokeLeaseByFileSystem(HadoopRecoverableFsDataOutputStream.java:327)\n\tat > > org.apache.flink.runtime.fs.hdfs.HadoopRecoverableFsDataOutputStream.safelyTruncateFile(HadoopRecoverableFsDataOutputStream.java:163)\n\tat > > org.apache.flink.runtime.fs.hdfs.HadoopRecoverableFsDataOutputStream.(HadoopRecoverableFsDataOutputStream.java:88)\n\tat > > org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.recover(HadoopRecoverableWriter.java:86)\n\tat > > org.apache.flink.streaming.api.functions.sink.filesystem.OutputStreamBasedPartFileWriter$OutputStreamBasedBucketWriter.resumeInProgressFileFrom(OutputStreamBasedPartFileWriter.java:104)\n\tat > org.apache.flink.streaming.api.functions.sink.filesyst > Task set 'execution.checkpointing.interval': 1min, I invoke savepoint > every fifth minutes. > Consult next everybody solution. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Created] (FLINK-25640) Enhance the document for blocking shuffle
Yingjie Cao created FLINK-25640: --- Summary: Enhance the document for blocking shuffle Key: FLINK-25640 URL: https://issues.apache.org/jira/browse/FLINK-25640 Project: Flink Issue Type: Sub-task Components: Documentation Reporter: Yingjie Cao Fix For: 1.15.0 As discussed in [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this ticket aims to enhance the document for blocking shuffle and add more operation guidelines. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * ffcee2854060f08f1f5ee3821f76823872010a15 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29359) * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lichang-bd commented on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
lichang-bd commented on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011863390 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-ml] lindong28 commented on pull request #55: [hotfix] Remove redundant directories
lindong28 commented on pull request #55: URL: https://github.com/apache/flink-ml/pull/55#issuecomment-1011863063 @gaoyunhaii @zhipeng93 Could you help review this PR? Thanks. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink-ml] lindong28 opened a new pull request #55: [hotfix] Remove redundant directories
lindong28 opened a new pull request #55: URL: https://github.com/apache/flink-ml/pull/55 ## What is the purpose of the change Remove redundant directories and make the directory structure consistent. ## Brief change log Moved `KnnTest.java` and `LogisticRegressionTest.java` to their parent directories. ## Verifying this change N/A ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (N/A) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-25639) Increase the default read buffer size of sort-shuffle to 64M
Yingjie Cao created FLINK-25639: --- Summary: Increase the default read buffer size of sort-shuffle to 64M Key: FLINK-25639 URL: https://issues.apache.org/jira/browse/FLINK-25639 Project: Flink Issue Type: Sub-task Components: Runtime / Network Reporter: Yingjie Cao Fix For: 1.15.0 As discussed in [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this ticket aims to increase the default read buffer size of sort-shuffle to 64M. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * ffcee2854060f08f1f5ee3821f76823872010a15 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29359) * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) * Unknown: [CANCELED](TBD) * 8337b385d6b96699e33fe7668d8b34bfc38beb2d UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18336: [FLINK-25441][network] Wrap failure cuase with ProducerFailedException only for PipelinedSubpartitionView
flinkbot edited a comment on pull request #18336: URL: https://github.com/apache/flink/pull/18336#issuecomment-1010777588 ## CI report: * 1a960b7f82e85d42e2a4f5eb9e9d63a203ed9de5 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29358) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-25559) SQL JOIN causes data loss
[ https://issues.apache.org/jira/browse/FLINK-25559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zongwen Li closed FLINK-25559. -- Fix Version/s: 1.15.0 1.13.6 1.14.3 Resolution: Fixed > SQL JOIN causes data loss > - > > Key: FLINK-25559 > URL: https://issues.apache.org/jira/browse/FLINK-25559 > Project: Flink > Issue Type: Bug > Components: Table SQL / API >Affects Versions: 1.14.0, 1.13.2, 1.13.3, 1.13.5, 1.14.2 >Reporter: Zongwen Li >Priority: Major > Fix For: 1.15.0, 1.13.6, 1.14.3 > > Attachments: image-2022-01-07-11-27-01-010.png > > > {code:java} > //sink table,omits some physical fields > CREATE TABLE kd_product_info ( > productId BIGINT COMMENT '产品编号', > productSaleId BIGINT COMMENT '商品编号', > PRIMARY KEY (productSaleId) NOT ENFORCED > ) > // sql omits some selected fields > INSERT INTO kd_product_info > SELECT > ps.product AS productId, > ps.productsaleid AS productSaleId, > CAST(p.complex AS INT) AS complex, > p.createtime AS createTime, > p.updatetime AS updateTime, > p.ean AS ean, > ts.availablequantity AS totalAvailableStock, > IF > ( > ts.availablequantity - ts.lockoccupy - ts.lock_available_quantity > 0, > ts.availablequantity - ts.lockoccupy - ts.lock_available_quantity, > 0 > ) AS sharedStock > ,rps.purchase AS purchase > ,v.`name` AS vendorName > FROM > product_sale ps > JOIN product p ON ps.product = p.id > LEFT JOIN rate_product_sale rps ON ps.productsaleid = rps.id > LEFT JOIN pss_total_stock ts ON ps.productsaleid = ts.productsale > LEFT JOIN vendor v ON ps.merchant_id = v.merchant_id AND ps.vendor = v.vendor > LEFT JOIN mccategory mc ON ps.merchant_id = mc.merchant_id AND p.mccategory > = mc.id > LEFT JOIN new_mccategory nmc ON p.mccategory = nmc.mc > LEFT JOIN product_sale_grade_plus psgp ON ps.productsaleid = psgp.productsale > LEFT JOIN product_sale_extend pse359 ON ps.productsaleid = > pse359.product_sale AND pse359.meta = 359 > LEFT JOIN product_image_url piu ON ps.product = piu.product {code} > All table sources are upsert-kafka,I have ensured that the associated columns > are of the same type: > {code:java} > CREATE TABLE product_sale ( > id BIGINT COMMENT '主键', > productsaleid BIGINT COMMENT '商品编号', > product BIGINT COMMENT '产品编号', > merchant_id DECIMAL(20, 0) COMMENT '商户id', > vendor STRING COMMENT '供应商', > PRIMARY KEY (productsaleid) NOT ENFORCED > ) > // No computed columns > // Just plain physical columns > WITH ( > 'connector' = 'upsert-kafka', > 'topic' = 'XXX', > 'group.id' = '%s', > 'properties.bootstrap.servers' = '%s', > 'key.format' = 'json', > 'value.format' = 'json' > ) > CREATE TABLE product ( > id BIGINT, > mccategory STRING, > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE rate_product_sale ( > id BIGINT COMMENT '主键', > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE pss_total_stock ( > id INT COMMENT 'ID', > productsale BIGINT COMMENT '商品编码', > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE vendor ( > merchant_id DECIMAL(20, 0) COMMENT '商户id', > vendor STRING COMMENT '供应商编码', > PRIMARY KEY (merchant_id, vendor) NOT ENFORCED > ) > CREATE TABLE mccategory ( > id STRING COMMENT 'mc编号', > merchant_id DECIMAL(20, 0) COMMENT '商户id', > PRIMARY KEY (merchant_id, id) NOT ENFORCED > ) > CREATE TABLE new_mccategory ( > mc STRING, > PRIMARY KEY (mc) NOT ENFORCED > ) > CREATE TABLE product_sale_grade_plus ( > productsale BIGINT, > PRIMARY KEY (productsale) NOT ENFORCED > ) > CREATE TABLE product_sale_extend ( > id BIGINT, > product_sale BIGINT, > meta BIGINT, > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE product_image_url( > product BIGINT, > PRIMARY KEY (product) NOT ENFORCED > ){code} > the data in each table is between 5 million and 10 million, parallelism: 24; > Not set ttl; In fact, we can notice data loss as soon as 30 minutes. > > The data flow: > MySQL -> Flink CDC -> ODS (Upsert Kafka) -> the job -> sink > I'm sure the ODS data in Kafka is correct; > I have also tried to use the flink-cdc source directly, it didn't solve the > problem; > > We tested sinking to kudu, Kafka or ES; > Also tested multiple versions: 1.13.2, 1.13.3, 1.13.5, 1.14.0, 1.14.2; > Lost data appears out of order on kafka, guessed as a bug of retraction > stream: > !image-2022-01-07-11-27-01-010.png! > > After many tests, we found that when the left join table is more or the > parallelism of the operator is greater, the data will be more easily lost. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] lichang-bd commented on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
lichang-bd commented on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011858591 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18215: [FLINK-25392][table-planner]Support new StatementSet syntax in planner and parser
flinkbot edited a comment on pull request #18215: URL: https://github.com/apache/flink/pull/18215#issuecomment-1001883080 ## CI report: * 795338a6386db7ede06778e819c48295463edc87 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29356) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] zlzhang0122 commented on pull request #18337: [hotfix][tests] Fix the error parameter
zlzhang0122 commented on pull request #18337: URL: https://github.com/apache/flink/pull/18337#issuecomment-1011857011 > Please open a ticket for issue. ok, I will create a ticket for this issue. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475136#comment-17475136 ] zlzhang0122 commented on FLINK-25634: - You can find the config yarn.application-attempt-failures-validity-interval in YarnConfigOptions, this is a flink conf. > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Commented] (FLINK-25615) FlinkKafkaProducer fail to correctly migrate pre Flink 1.9 state
[ https://issues.apache.org/jira/browse/FLINK-25615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475139#comment-17475139 ] Matthias Schwalbe commented on FLINK-25615: --- Thought so :) However, it is documented should someone run into the problem, which was not easy to locate. ... next time then :) Happy Flinking Thias > FlinkKafkaProducer fail to correctly migrate pre Flink 1.9 state > > > Key: FLINK-25615 > URL: https://issues.apache.org/jira/browse/FLINK-25615 > Project: Flink > Issue Type: Bug > Components: Connectors / Kafka >Affects Versions: 1.9.0 >Reporter: Matthias Schwalbe >Priority: Major > > I've found an unnoticed error in FlinkKafkaProvider when migrating from pre > Flink 1.9 state to versions starting with Flink 1.9: > * the operator state for next-transactional-id-hint should be deleted and > replaced by operator state next-transactional-id-hint-v2, however > * operator state next-transactional-id-hint is never deleted > * see here: [1] : > {quote} if (context.getOperatorStateStore() > .getRegisteredStateNames() > .contains(NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR)) { > migrateNextTransactionalIdHindState(context); > }{quote} * migrateNextTransactionalIdHindState is never called, as > the condition cannot become true: > ** getRegisteredStateNames returns a list of String, whereas > NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR is ListStateDescriptor (type mismatch) > The Effect is: > * because NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR is for a UnionListState, and > * the state is not cleared, > * each time the job restarts from a savepoint or checkpoint the size > multiplies times the parallelism > * then because each entry leaves an offset in metadata, akka.framesize > becomes too small, before we run into memory overflow > > The breaking change has been introduced in commit > 70fa80e3862b367be22b593db685f9898a2838ef > > A simple fix would be to change the code to: > {quote} if (context.getOperatorStateStore() > .getRegisteredStateNames() > .contains(NEXT_TRANSACTIONAL_ID_HINT_DESCRIPTOR.getName())) { > migrateNextTransactionalIdHindState(context); > } > {quote} > > Although FlinkKafkaProvider is marked as deprecated it is probably a while > here to stay > > Greeting > Matthias (Thias) Schwalbe > > [1] > https://github.com/apache/flink/blob/d7cf2c10f8d4fba81173854cbd8be27c657c7c7f/flink-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaProducer.java#L1167-L1171 > -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] BruceWong96 edited a comment on pull request #18293: [FLINK-25560][Connectors/HBase] Add "sink.delete.mode" in HBase sql connector for retracting the latest version or all versions in
BruceWong96 edited a comment on pull request #18293: URL: https://github.com/apache/flink/pull/18293#issuecomment-1009895174 > Thanks for the contribution. The change should be working for your case and the code itself looks overall ok. It looks like you want to fix the HBase bug by adding work-around into the connector and I have following concerns: > > 1. Extra customized config will be introduced to users which turns out that an internal bug effects users' behaviour. > 2. The HBaseConnectorOptions has been extended for specific process logic, beyond the neutral HBase Lookup Options, HBase Configuration Options, and Flink Operator Options. This smells like the start of the hacking. > 3. HBase should be the right one to control the versions. With the change of this PR, there are two roles who can do the same job, a SRP violation. It should be used very carefully to avoid data lost in the production. I could already imagine the situation where user would complain the unexpected data lost. I understood that the default deleteMode is trying to reduce such incident, thanks for considering about that, but it is now possible for e.g. human mistake on the SQL level to make the data lost happen and, once it happens, it is expensive in the production env. > 4. This change does not cover multiple CF scenario, since the scope of version has been changed/enlarged from CF to table. Modifying the deleteMode for one CF might end up with the data lost of another CF. if not modify it, it will still have the deletion issue for the related CF. This could be solved by splitting the table, which turns out again an internal bug effects users' behaviour. There will be extra effort to maintain two maybe even more tables that conceptually belong together and keep the data synched. > 5. Shotgun surgery - too many classes have been touched for building the work-around of simple functionality. > 6. Classes defined for common case got involved into specific deletion case, i.e. HBaseWriteOptions, HBaseMutationConverter, etc. > > Looking forward to your thoughts. Thank you for your reply 1. Users need to understand the usage of the custom configuration before using it. 2. In change log mode, op= 'd' means deleting data, even before this PR. 3. In change log mode, Flink SQL calls HBase Client to modify data, which does not violate the single responsibility principle. Flink SQL can also delete data before this PR. If the user selects' sink.delete.mode '=' all-versions', the data is not accidentally lost but deleted normally. Before the PR, only the last version is deleted. 4. After the test, we found that the deleted CF was the CF specified in the Flink SQL DML statement, without affecting other CF. The specific test results can be found in JIRA. 5. I think it is not a good way to avoid data loss without providing a correct way to delete data. What we need is that the HBase Connector can meet the requirements of adding, deleting, modifying and querying in the change log mode. It seems that the deletion is not perfect, which leads to incorrect data. Looking forward to your thoughts. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-25631) Support enhanced `show tables` statement
[ https://issues.apache.org/jira/browse/FLINK-25631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475138#comment-17475138 ] Martijn Visser commented on FLINK-25631: Thanks [~jark] and [~liyubin117]! > Support enhanced `show tables` statement > > > Key: FLINK-25631 > URL: https://issues.apache.org/jira/browse/FLINK-25631 > Project: Flink > Issue Type: New Feature > Components: Table SQL / API >Affects Versions: 1.14.4 >Reporter: Yubin Li >Assignee: Yubin Li >Priority: Major > > Enhanced `show tables` statement like ` show tables from db1 like 't%' ` has > been supported broadly in many popular data process engine like > presto/trino/spark > [https://spark.apache.org/docs/latest/sql-ref-syntax-aux-show-tables.html] > I have investigated the syntax of engines as mentioned above. > > We could use such statement to easily show the tables of specified databases > without switching db frequently, alse we could use regexp pattern to find > focused tables quickly from plenty of tables. besides, the new statement is > compatible completely with the old one, users could use `show tables` as > before. > h3. SHOW TABLES [ ( FROM | IN ) [catalog.]db ] [ [NOT] LIKE regex_pattern ] > > I have implemented the syntax, demo as below: > {code:java} > Flink SQL> create database d1; > [INFO] Execute statement succeed. > Flink SQL> create table d1.b1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> create table t1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> create table m1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> show tables like 'm%'; > ++ > | table name | > ++ > | m1 | > ++ > 1 row in set > Flink SQL> show tables from d1 like 'b%'; > ++ > | table name | > ++ > | b1 | > ++ > 1 row in set{code} -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Commented] (FLINK-25559) SQL JOIN causes data loss
[ https://issues.apache.org/jira/browse/FLINK-25559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475137#comment-17475137 ] Zongwen Li commented on FLINK-25559: [~monster#12] Following Lincoln's suggestion, I tried using flink 1.14.3-rc1 and so far the problem has not recurred. > SQL JOIN causes data loss > - > > Key: FLINK-25559 > URL: https://issues.apache.org/jira/browse/FLINK-25559 > Project: Flink > Issue Type: Bug > Components: Table SQL / API >Affects Versions: 1.14.0, 1.13.2, 1.13.3, 1.13.5, 1.14.2 >Reporter: Zongwen Li >Priority: Major > Attachments: image-2022-01-07-11-27-01-010.png > > > {code:java} > //sink table,omits some physical fields > CREATE TABLE kd_product_info ( > productId BIGINT COMMENT '产品编号', > productSaleId BIGINT COMMENT '商品编号', > PRIMARY KEY (productSaleId) NOT ENFORCED > ) > // sql omits some selected fields > INSERT INTO kd_product_info > SELECT > ps.product AS productId, > ps.productsaleid AS productSaleId, > CAST(p.complex AS INT) AS complex, > p.createtime AS createTime, > p.updatetime AS updateTime, > p.ean AS ean, > ts.availablequantity AS totalAvailableStock, > IF > ( > ts.availablequantity - ts.lockoccupy - ts.lock_available_quantity > 0, > ts.availablequantity - ts.lockoccupy - ts.lock_available_quantity, > 0 > ) AS sharedStock > ,rps.purchase AS purchase > ,v.`name` AS vendorName > FROM > product_sale ps > JOIN product p ON ps.product = p.id > LEFT JOIN rate_product_sale rps ON ps.productsaleid = rps.id > LEFT JOIN pss_total_stock ts ON ps.productsaleid = ts.productsale > LEFT JOIN vendor v ON ps.merchant_id = v.merchant_id AND ps.vendor = v.vendor > LEFT JOIN mccategory mc ON ps.merchant_id = mc.merchant_id AND p.mccategory > = mc.id > LEFT JOIN new_mccategory nmc ON p.mccategory = nmc.mc > LEFT JOIN product_sale_grade_plus psgp ON ps.productsaleid = psgp.productsale > LEFT JOIN product_sale_extend pse359 ON ps.productsaleid = > pse359.product_sale AND pse359.meta = 359 > LEFT JOIN product_image_url piu ON ps.product = piu.product {code} > All table sources are upsert-kafka,I have ensured that the associated columns > are of the same type: > {code:java} > CREATE TABLE product_sale ( > id BIGINT COMMENT '主键', > productsaleid BIGINT COMMENT '商品编号', > product BIGINT COMMENT '产品编号', > merchant_id DECIMAL(20, 0) COMMENT '商户id', > vendor STRING COMMENT '供应商', > PRIMARY KEY (productsaleid) NOT ENFORCED > ) > // No computed columns > // Just plain physical columns > WITH ( > 'connector' = 'upsert-kafka', > 'topic' = 'XXX', > 'group.id' = '%s', > 'properties.bootstrap.servers' = '%s', > 'key.format' = 'json', > 'value.format' = 'json' > ) > CREATE TABLE product ( > id BIGINT, > mccategory STRING, > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE rate_product_sale ( > id BIGINT COMMENT '主键', > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE pss_total_stock ( > id INT COMMENT 'ID', > productsale BIGINT COMMENT '商品编码', > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE vendor ( > merchant_id DECIMAL(20, 0) COMMENT '商户id', > vendor STRING COMMENT '供应商编码', > PRIMARY KEY (merchant_id, vendor) NOT ENFORCED > ) > CREATE TABLE mccategory ( > id STRING COMMENT 'mc编号', > merchant_id DECIMAL(20, 0) COMMENT '商户id', > PRIMARY KEY (merchant_id, id) NOT ENFORCED > ) > CREATE TABLE new_mccategory ( > mc STRING, > PRIMARY KEY (mc) NOT ENFORCED > ) > CREATE TABLE product_sale_grade_plus ( > productsale BIGINT, > PRIMARY KEY (productsale) NOT ENFORCED > ) > CREATE TABLE product_sale_extend ( > id BIGINT, > product_sale BIGINT, > meta BIGINT, > PRIMARY KEY (id) NOT ENFORCED > ) > CREATE TABLE product_image_url( > product BIGINT, > PRIMARY KEY (product) NOT ENFORCED > ){code} > the data in each table is between 5 million and 10 million, parallelism: 24; > Not set ttl; In fact, we can notice data loss as soon as 30 minutes. > > The data flow: > MySQL -> Flink CDC -> ODS (Upsert Kafka) -> the job -> sink > I'm sure the ODS data in Kafka is correct; > I have also tried to use the flink-cdc source directly, it didn't solve the > problem; > > We tested sinking to kudu, Kafka or ES; > Also tested multiple versions: 1.13.2, 1.13.3, 1.13.5, 1.14.0, 1.14.2; > Lost data appears out of order on kafka, guessed as a bug of retraction > stream: > !image-2022-01-07-11-27-01-010.png! > > After many tests, we found that when the left join table is more or the > parallelism of the operator is greater, the data will be more easily lost. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #17676: [FLINK-14100][connectors] Add Oracle dialect
flinkbot edited a comment on pull request #17676: URL: https://github.com/apache/flink/pull/17676#issuecomment-960641444 ## CI report: * 35238e2baa5a30a5a5e047aceddedbfae6cfcd75 Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29357) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Comment Edited] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475136#comment-17475136 ] zlzhang0122 edited comment on FLINK-25634 at 1/13/22, 7:04 AM: --- You can find the config yarn.application-attempt-failures-validity-interval in [YarnConfigOptions|https://github.com/apache/flink/blob/dbbf2a36111da1faea5c901e3b008cc788913bf8/flink-yarn/src/main/java/org/apache/flink/yarn/configuration/YarnConfigOptions.java#L106], this is a flink conf. was (Author: zlzhang0122): You can find the config yarn.application-attempt-failures-validity-interval in YarnConfigOptions, this is a flink conf. > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * ffcee2854060f08f1f5ee3821f76823872010a15 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29359) * 86046d08986f3ed691c9766d104f97163c445192 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29364) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Comment Edited] (FLINK-25484) TableRollingPolicy do not support inactivityInterval config which is supported in datastream api
[ https://issues.apache.org/jira/browse/FLINK-25484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475088#comment-17475088 ] LiChang edited comment on FLINK-25484 at 1/13/22, 7:02 AM: --- yes,I am doing on it [~fpaul] you may assigned this task to me was (Author: lichang): yes,I am doing on it Fabian Paul > TableRollingPolicy do not support inactivityInterval config which is > supported in datastream api > > > Key: FLINK-25484 > URL: https://issues.apache.org/jira/browse/FLINK-25484 > Project: Flink > Issue Type: Improvement > Components: Connectors / FileSystem, Table SQL / Ecosystem >Affects Versions: 1.15.0 >Reporter: LiChang >Priority: Major > Fix For: 1.15.0 > > Attachments: image-2022-01-13-11-36-24-102.png > > > TableRollingPolicy do not support inactivityInterval config > public static class TableRollingPolicy extends > CheckpointRollingPolicy { > private final boolean rollOnCheckpoint; > private final long rollingFileSize; > private final long rollingTimeInterval; > public TableRollingPolicy( > boolean rollOnCheckpoint, long rollingFileSize, long rollingTimeInterval) { > this.rollOnCheckpoint = rollOnCheckpoint; > Preconditions.checkArgument(rollingFileSize > 0L); > Preconditions.checkArgument(rollingTimeInterval > 0L); > this.rollingFileSize = rollingFileSize; > this.rollingTimeInterval = rollingTimeInterval; > } > -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Created] (FLINK-25638) Increase the default write buffer size of sort-shuffle to 16M
Yingjie Cao created FLINK-25638: --- Summary: Increase the default write buffer size of sort-shuffle to 16M Key: FLINK-25638 URL: https://issues.apache.org/jira/browse/FLINK-25638 Project: Flink Issue Type: Sub-task Components: Runtime / Network Reporter: Yingjie Cao Fix For: 1.15.0 As discussed in [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this ticket aims to increase the default write buffer size of sort-shuffle to 16M. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Created] (FLINK-25637) Make sort-shuffle the default shuffle implementation for batch jobs
Yingjie Cao created FLINK-25637: --- Summary: Make sort-shuffle the default shuffle implementation for batch jobs Key: FLINK-25637 URL: https://issues.apache.org/jira/browse/FLINK-25637 Project: Flink Issue Type: Sub-task Components: Runtime / Network Reporter: Yingjie Cao Fix For: 1.15.0 As discussed in [https://lists.apache.org/thread/pt2b1f17x2l5rlvggwxs6m265lo4ly7p], this ticket aims to make sort-shuffle the default shuffle implementation for batch jobs. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Commented] (FLINK-25631) Support enhanced `show tables` statement
[ https://issues.apache.org/jira/browse/FLINK-25631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475133#comment-17475133 ] Jark Wu commented on FLINK-25631: - Thanks [~liyubin117], this feature sounds good to me. You can take FLINK-22885 as an example to implement it (including the tests and docs). > Support enhanced `show tables` statement > > > Key: FLINK-25631 > URL: https://issues.apache.org/jira/browse/FLINK-25631 > Project: Flink > Issue Type: New Feature > Components: Table SQL / API >Affects Versions: 1.14.4 >Reporter: Yubin Li >Assignee: Yubin Li >Priority: Major > > Enhanced `show tables` statement like ` show tables from db1 like 't%' ` has > been supported broadly in many popular data process engine like > presto/trino/spark > [https://spark.apache.org/docs/latest/sql-ref-syntax-aux-show-tables.html] > I have investigated the syntax of engines as mentioned above. > > We could use such statement to easily show the tables of specified databases > without switching db frequently, alse we could use regexp pattern to find > focused tables quickly from plenty of tables. besides, the new statement is > compatible completely with the old one, users could use `show tables` as > before. > h3. SHOW TABLES [ ( FROM | IN ) [catalog.]db ] [ [NOT] LIKE regex_pattern ] > > I have implemented the syntax, demo as below: > {code:java} > Flink SQL> create database d1; > [INFO] Execute statement succeed. > Flink SQL> create table d1.b1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> create table t1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> create table m1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> show tables like 'm%'; > ++ > | table name | > ++ > | m1 | > ++ > 1 row in set > Flink SQL> show tables from d1 like 'b%'; > ++ > | table name | > ++ > | b1 | > ++ > 1 row in set{code} -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * ffcee2854060f08f1f5ee3821f76823872010a15 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29359) * 86046d08986f3ed691c9766d104f97163c445192 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-25631) Support enhanced `show tables` statement
[ https://issues.apache.org/jira/browse/FLINK-25631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu reassigned FLINK-25631: --- Assignee: Yubin Li > Support enhanced `show tables` statement > > > Key: FLINK-25631 > URL: https://issues.apache.org/jira/browse/FLINK-25631 > Project: Flink > Issue Type: New Feature > Components: Table SQL / API >Affects Versions: 1.14.4 >Reporter: Yubin Li >Assignee: Yubin Li >Priority: Major > > Enhanced `show tables` statement like ` show tables from db1 like 't%' ` has > been supported broadly in many popular data process engine like > presto/trino/spark > [https://spark.apache.org/docs/latest/sql-ref-syntax-aux-show-tables.html] > I have investigated the syntax of engines as mentioned above. > > We could use such statement to easily show the tables of specified databases > without switching db frequently, alse we could use regexp pattern to find > focused tables quickly from plenty of tables. besides, the new statement is > compatible completely with the old one, users could use `show tables` as > before. > h3. SHOW TABLES [ ( FROM | IN ) [catalog.]db ] [ [NOT] LIKE regex_pattern ] > > I have implemented the syntax, demo as below: > {code:java} > Flink SQL> create database d1; > [INFO] Execute statement succeed. > Flink SQL> create table d1.b1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> create table t1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> create table m1(id int) with ('connector'='print'); > [INFO] Execute statement succeed. > Flink SQL> show tables like 'm%'; > ++ > | table name | > ++ > | m1 | > ++ > 1 row in set > Flink SQL> show tables from d1 like 'b%'; > ++ > | table name | > ++ > | b1 | > ++ > 1 row in set{code} -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] wsry commented on a change in pull request #17814: [FLINK-24899][runtime] Enable data compression for blocking shuffle by default
wsry commented on a change in pull request #17814: URL: https://github.com/apache/flink/pull/17814#discussion_r783667878 ## File path: flink-tests/src/test/java/org/apache/flink/test/runtime/ShuffleCompressionITCase.java ## @@ -84,8 +84,6 @@ @Test public void testDataCompressionForBoundedBlockingShuffle() throws Exception { Review comment: After changing the default value to true, I guess many test cases including BlockingShuffleITCase have already covered the scenario of data compression. Here what we really need is a test case which covers the scenario where data is not compressed. Could you please change the method name including the below one to something like testBoundedBlockingShuffleWithoutCompression/testSortMergeBlockingShuffleWithoutCompression, and set the bool flag of data compression to false? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry commented on pull request #17814: [FLINK-24899][runtime] Enable data compression for blocking shuffle by default
wsry commented on pull request #17814: URL: https://github.com/apache/flink/pull/17814#issuecomment-1011848043 @SteNicholas Thanks for your contribution, I have left a small comment. BTW, could you please also update the corresponding document for blocking shuffle? (docs/content/docs/ops/batch/blocking_shuffle.md & docs/content.zh/docs/ops/batch/blocking_shuffle.md) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18201: [FLINK-25167][API / DataStream]Expose StreamOperatorFactory in Connec…
flinkbot edited a comment on pull request #18201: URL: https://github.com/apache/flink/pull/18201#issuecomment-1000767373 ## CI report: * ae038dbc84bb1cc7d6654e40081c01b9c1211411 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29355) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18346: [DO NOT MERGE][FLINK-25085] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
flinkbot edited a comment on pull request #18346: URL: https://github.com/apache/flink/pull/18346#issuecomment-1011842175 ## CI report: * 195e75ed899f947fbb8b535f9bfc16f501d0f86e Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29363) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634 ] 宇宙先生 deleted comment on FLINK-25634: -- was (Author: JIRAUSER283011): firstly,thanks for your answer,But I can't find this config yarn.application-attempt-failures-validity-interval,can you help me check it > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot commented on pull request #18346: [DO NOT MERGE][FLINK-25085] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
flinkbot commented on pull request #18346: URL: https://github.com/apache/flink/pull/18346#issuecomment-1011842594 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 195e75ed899f947fbb8b535f9bfc16f501d0f86e (Thu Jan 13 06:44:11 UTC 2022) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #18346: [DO NOT MERGE][FLINK-25085] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
flinkbot commented on pull request #18346: URL: https://github.com/apache/flink/pull/18346#issuecomment-1011842175 ## CI report: * 195e75ed899f947fbb8b535f9bfc16f501d0f86e UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] zjureel opened a new pull request #18346: [DO NOT MERGE] Add scheduled executor service in MainThreadExecutor and close it when the server reaches termination
zjureel opened a new pull request #18346: URL: https://github.com/apache/flink/pull/18346 ## What is the purpose of the change *(For example: This pull request makes task deployment go through the blob server, rather than through RPC. That way we avoid re-transferring them on each deployment (during recovery).)* ## Brief change log *(for example:)* - *The TaskInfo is stored in the blob store on job creation time as a persistent artifact* - *Deployments RPC transmits only the blob storage reference* - *TaskManagers retrieve the TaskInfo from the blob cache* ## Verifying this change Please make sure both new and modified tests in this PR follows the conventions defined in our code quality guide: https://flink.apache.org/contributing/code-style-and-quality-common.html#testing *(Please pick either of the following options)* This change is a trivial rework / code cleanup without any test coverage. *(or)* This change is already covered by existing tests, such as *(please describe tests)*. *(or)* This change added tests and can be verified as follows: *(example:)* - *Added integration tests for end-to-end deployment with large payloads (100MB)* - *Extended integration test for recovery after master (JobManager) failure* - *Added test that validates that TaskInfo is transferred only once across recoveries* - *Manually verified the change by running a 4 node cluser with 2 JobManagers and 4 TaskManagers, a stateful streaming program, and killing one JobManager and two TaskManagers during the execution, verifying that recovery happens correctly.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / no) - The serializers: (yes / no / don't know) - The runtime per-record code paths (performance sensitive): (yes / no / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know) - The S3 file system connector: (yes / no / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / no) - If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475127#comment-17475127 ] 宇宙先生 commented on FLINK-25634: -- firstly,thanks for your answer,But I can't find this config yarn.application-attempt-failures-validity-interval,can you help me check it > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] shouweikun commented on pull request #18201: [FLINK-25167][API / DataStream]Expose StreamOperatorFactory in Connec…
shouweikun commented on pull request #18201: URL: https://github.com/apache/flink/pull/18201#issuecomment-1011838739 Hi, @fapaul @pnowojski I‘d say that I encountered some problems. I tried to mark `StreamOperatorFactory` and `StreamOperatorParameters` as `@PublicEvoliving`, tests still failed. I checked the cause and intended to fix them but I didn't know what to do. Asking for help. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-24899) Enable data compression for blocking shuffle by default
[ https://issues.apache.org/jira/browse/FLINK-24899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao reassigned FLINK-24899: --- Assignee: Nicholas Jiang > Enable data compression for blocking shuffle by default > --- > > Key: FLINK-24899 > URL: https://issues.apache.org/jira/browse/FLINK-24899 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Network >Reporter: Yingjie Cao >Assignee: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: 1.15.0 > > > Currently, shuffle data compression is not enabled by default. Shuffle data > compression is important for blocking data shuffle and enabling shuffle data > compression by default can improve the usability. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634 ] 宇宙先生 deleted comment on FLINK-25634: -- was (Author: JIRAUSER283011): first,thanks for your answer,but I can't find this config yarn.application-attempt-failures-validity-interval , can you help me check it ,thanks a lot! https://hadoop.apache.org/docs/r2.4.1/hadoop-yarn/hadoop-yarn-common/yarn-default.xml !image-2022-01-13-14-19-43-539.png! > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Updated] (FLINK-24899) Enable data compression for blocking shuffle by default
[ https://issues.apache.org/jira/browse/FLINK-24899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao updated FLINK-24899: Parent: (was: FLINK-24898) Issue Type: Improvement (was: Sub-task) > Enable data compression for blocking shuffle by default > --- > > Key: FLINK-24899 > URL: https://issues.apache.org/jira/browse/FLINK-24899 > Project: Flink > Issue Type: Improvement > Components: Runtime / Network >Reporter: Yingjie Cao >Priority: Major > Labels: pull-request-available > Fix For: 1.15.0 > > > Currently, shuffle data compression is not enabled by default. Shuffle data > compression is important for blocking data shuffle and enabling shuffle data > compression by default can improve the usability. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Updated] (FLINK-24899) Enable data compression for blocking shuffle by default
[ https://issues.apache.org/jira/browse/FLINK-24899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao updated FLINK-24899: Parent: FLINK-25636 Issue Type: Sub-task (was: Improvement) > Enable data compression for blocking shuffle by default > --- > > Key: FLINK-24899 > URL: https://issues.apache.org/jira/browse/FLINK-24899 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Network >Reporter: Yingjie Cao >Priority: Major > Labels: pull-request-available > Fix For: 1.15.0 > > > Currently, shuffle data compression is not enabled by default. Shuffle data > compression is important for blocking data shuffle and enabling shuffle data > compression by default can improve the usability. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Created] (FLINK-25636) FLIP-199: Change some default config values of blocking shuffle for better usability
Yingjie Cao created FLINK-25636: --- Summary: FLIP-199: Change some default config values of blocking shuffle for better usability Key: FLINK-25636 URL: https://issues.apache.org/jira/browse/FLINK-25636 Project: Flink Issue Type: Improvement Components: Runtime / Network Reporter: Yingjie Cao Fix For: 1.15.0 This is the umbrella issue for FLIP-199, we will change the several default config value for batch shuffle and update the document accordingly. -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Commented] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17475124#comment-17475124 ] 宇宙先生 commented on FLINK-25634: -- first,thanks for your answer,but I can't find this config yarn.application-attempt-failures-validity-interval , can you help me check it ,thanks a lot! https://hadoop.apache.org/docs/r2.4.1/hadoop-yarn/hadoop-yarn-common/yarn-default.xml !image-2022-01-13-14-19-43-539.png! > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[jira] [Updated] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] 宇宙先生 updated FLINK-25634: - Attachment: image-2022-01-13-14-19-43-539.png > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png, image-2022-01-13-14-19-43-539.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #17936: [FLINK-24954][network] Refresh read buffer request timeout on buffer recycling/requesting for sort-shuffle
flinkbot edited a comment on pull request #17936: URL: https://github.com/apache/flink/pull/17936#issuecomment-980533423 ## CI report: * 1ec81b2f7c5abbf5a2bd470dc804977085787e4c Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29303) * 52013eb761deb1f972813d863269ae959fa11267 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29362) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry commented on pull request #17936: [FLINK-24954][network] Refresh read buffer request timeout on buffer recycling/requesting for sort-shuffle
wsry commented on pull request #17936: URL: https://github.com/apache/flink/pull/17936#issuecomment-1011830027 @TanYuxin-tyx Thanks for the contribution, changes LGTM. I will merge it after build succeeds. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #17936: [FLINK-24954][network] Refresh read buffer request timeout on buffer recycling/requesting for sort-shuffle
flinkbot edited a comment on pull request #17936: URL: https://github.com/apache/flink/pull/17936#issuecomment-980533423 ## CI report: * 1ec81b2f7c5abbf5a2bd470dc804977085787e4c Azure: [SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29303) * 52013eb761deb1f972813d863269ae959fa11267 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] imaffe commented on pull request #17937: [FLINK-25044][testing][Pulsar Connector] Add More Unit Test For Pulsar Source
imaffe commented on pull request #17937: URL: https://github.com/apache/flink/pull/17937#issuecomment-1011826129 > I think two comments are not addressed but everything else looks good now, thanks for the patience. Can you squash all the commits together that the PR only includes two commits one for the bug fix and one for the tests? Nit: Next time it would be great if you mark the comments you have addressed :) The first case can be deleted. The edge case it tries to test is addressed in the SplitReaderTest. Basically what we want to do here is to wait for a certain timeout between two messages in the pulsar client and see how it behaves. Do we have suggestions on how to test these kinds of behaviour? (needs to wait for an external system to timeout, feels we can't use a blocking queue to signal a event complete here). The second place, I'll set the polling timeout to a large number. And rely on the global timeout to terminate execution if the end condition is not met. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-25634) flink-read-onyarn-configuration
[ https://issues.apache.org/jira/browse/FLINK-25634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] 宇宙先生 updated FLINK-25634: - Description: in flink-src.code : !image-2022-01-13-10-14-18-918.png! Set the number of retries for failed YARN ApplicationMasters/JobManagers in high availability mode. This value is usually limited by YARN. By default, it's 1 in the standalone case and 2 in the high availability case in my cluster,the number of retries for failed YARN ApplicationMasters is 2 yarn's configuration also like this !image-2022-01-13-10-07-02-908.png! But it keeps restarting when my task fails, !image-2022-01-13-10-10-44-945.png! I would like to know the reason why the configuration is not taking effect. sincere thanks! was: in flink-src.code : !image-2022-01-13-10-14-18-918.png! Set the number of retries for failed YARN ApplicationMasters/JobManagers in high availability mode. This value is usually limited by YARN. By default, it's 1 in the standalone case and 2 in the high availability case in my cluster,the number of retries for failed YARN ApplicationMasters is 2 yarn's configuration also like this !image-2022-01-13-10-07-02-908.png! But it keeps restarting when my taskfails, !image-2022-01-13-10-10-44-945.png! I would like to know the reason why the configuration is not taking effect. sincere thanks! > flink-read-onyarn-configuration > --- > > Key: FLINK-25634 > URL: https://issues.apache.org/jira/browse/FLINK-25634 > Project: Flink > Issue Type: Bug >Affects Versions: 1.11.2 >Reporter: 宇宙先生 >Priority: Major > Attachments: image-2022-01-13-10-02-57-803.png, > image-2022-01-13-10-03-28-230.png, image-2022-01-13-10-07-02-908.png, > image-2022-01-13-10-09-36-890.png, image-2022-01-13-10-10-44-945.png, > image-2022-01-13-10-14-18-918.png > > > in flink-src.code : > !image-2022-01-13-10-14-18-918.png! > Set the number of retries for failed YARN ApplicationMasters/JobManagers in > high > availability mode. This value is usually limited by YARN. > By default, it's 1 in the standalone case and 2 in the high availability case > > in my cluster,the number of retries for failed YARN ApplicationMasters is 2 > yarn's configuration also like this > !image-2022-01-13-10-07-02-908.png! > But it keeps restarting when my task fails, > !image-2022-01-13-10-10-44-945.png! > I would like to know the reason why the configuration is not taking effect. > sincere thanks! -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] flinkbot edited a comment on pull request #18345: [FLINK-25440][doc] startCurosr only supports publishTime not eventTime
flinkbot edited a comment on pull request #18345: URL: https://github.com/apache/flink/pull/18345#issuecomment-1011821419 ## CI report: * 1dbd224db334939fc5337ac994d1b24efdf99657 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29361) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #18345: [FLINK-25440][doc] startCurosr only supports publishTime not eventTime
flinkbot commented on pull request #18345: URL: https://github.com/apache/flink/pull/18345#issuecomment-1011821419 ## CI report: * 1dbd224db334939fc5337ac994d1b24efdf99657 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #18345: [FLINK-25440][doc] startCurosr only supports publishTime not eventTime
flinkbot commented on pull request #18345: URL: https://github.com/apache/flink/pull/18345#issuecomment-1011821142 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit 1dbd224db334939fc5337ac994d1b24efdf99657 (Thu Jan 13 05:58:27 UTC 2022) **Warnings:** * Documentation files were touched, but no `docs/content.zh/` files: Update Chinese documentation or file Jira ticket. * **This pull request references an unassigned [Jira ticket](https://issues.apache.org/jira/browse/FLINK-25440).** According to the [code contribution guide](https://flink.apache.org/contributing/contribute-code.html), tickets need to be assigned before starting with the implementation work. Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] springMoon removed a comment on pull request #18309: [FLINK-25582] [TABLE SQL / API] flink sql kafka source cannot special custom parallelism
springMoon removed a comment on pull request #18309: URL: https://github.com/apache/flink/pull/18309#issuecomment-1011819956 @alpreu thanks for review, but @lmagic233 is right, generic solution for custom source parallelism is better -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] springMoon closed pull request #18309: [FLINK-25582] [TABLE SQL / API] flink sql kafka source cannot special custom parallelism
springMoon closed pull request #18309: URL: https://github.com/apache/flink/pull/18309 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] springMoon commented on pull request #18309: [FLINK-25582] [TABLE SQL / API] flink sql kafka source cannot special custom parallelism
springMoon commented on pull request #18309: URL: https://github.com/apache/flink/pull/18309#issuecomment-1011819956 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-25440) Apache Pulsar Connector Document description error about 'Starting Position'.
[ https://issues.apache.org/jira/browse/FLINK-25440?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-25440: --- Labels: pull-request-available (was: ) > Apache Pulsar Connector Document description error about 'Starting Position'. > - > > Key: FLINK-25440 > URL: https://issues.apache.org/jira/browse/FLINK-25440 > Project: Flink > Issue Type: Bug > Components: Documentation >Affects Versions: 1.14.2 >Reporter: xiechenling >Priority: Minor > Labels: pull-request-available > > Starting Position description error. > Start from the specified message time by Message.getEventTime(). > StartCursor.fromMessageTime(long) > it should be 'Start from the specified message time by publishTime.' -- This message was sent by Atlassian Jira (v8.20.1#820001)
[GitHub] [flink] imaffe opened a new pull request #18345: [FLINK-25440][doc] startCurosr only supports publishTime not eventTime
imaffe opened a new pull request #18345: URL: https://github.com/apache/flink/pull/18345
[GitHub] [flink] flinkbot edited a comment on pull request #17777: [FLINK-24886][core] TimeUtils supports the form of m.
flinkbot edited a comment on pull request #1: URL: https://github.com/apache/flink/pull/1#issuecomment-966971402 ## CI report: * b00b49e806baf132df4f13124b98469cef9ebb91 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29288) * 34f5bb8e51b6a57f934ff58d53eeb87252911c84 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29360) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] springMoon commented on pull request #18309: [FLINK-25582] [TABLE SQL / API] flink sql kafka source cannot special custom parallelism
springMoon commented on pull request #18309: URL: https://github.com/apache/flink/pull/18309#issuecomment-1011819020 @alpreu thanks for review, but @lmagic233 is right, generic solution for custom source parallelism is better -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #17777: [FLINK-24886][core] TimeUtils supports the form of m.
flinkbot edited a comment on pull request #1: URL: https://github.com/apache/flink/pull/1#issuecomment-966971402 ## CI report: * b00b49e806baf132df4f13124b98469cef9ebb91 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29288) * 34f5bb8e51b6a57f934ff58d53eeb87252911c84 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] liuyongvs commented on pull request #17777: [FLINK-24886][core] TimeUtils supports the form of m.
liuyongvs commented on pull request #1: URL: https://github.com/apache/flink/pull/1#issuecomment-1011814363 > ``` > MINUTES(ChronoUnit.MINUTES, singular("min"), singular("m"), plural("minute")) > ``` yes, but if we change to MINUTES(ChronoUnit.MINUTES, singular("min"), singular("m"), plural("minute")). i worry the user later may be weird。 if you think it is ok, i will change the order. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry closed pull request #18235: Test change default batch configuration
wsry closed pull request #18235: URL: https://github.com/apache/flink/pull/18235 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot edited a comment on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * ffcee2854060f08f1f5ee3821f76823872010a15 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29359) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot commented on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809921 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Automated Checks Last check on commit ffcee2854060f08f1f5ee3821f76823872010a15 (Thu Jan 13 05:30:15 UTC 2022) **Warnings:** * No documentation files were touched! Remember to keep the Flink docs up to date! * **This pull request references an unassigned [Jira ticket](https://issues.apache.org/jira/browse/FLINK-25484).** According to the [code contribution guide](https://flink.apache.org/contributing/contribute-code.html), tickets need to be assigned before starting with the implementation work. Mention the bot in a comment to re-run the automated checks. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
flinkbot commented on pull request #18344: URL: https://github.com/apache/flink/pull/18344#issuecomment-1011809365 ## CI report: * ffcee2854060f08f1f5ee3821f76823872010a15 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lichang-bd opened a new pull request #18344: [Flink-25484][Filesystem] Support inactivityInterval config in table api
lichang-bd opened a new pull request #18344: URL: https://github.com/apache/flink/pull/18344 Change-Id: Ica8ba43307476198c4f525ac40a54ab862b58ea7 ## What is the purpose of the change *(For example: This pull request makes task deployment go through the blob server, rather than through RPC. That way we avoid re-transferring them on each deployment (during recovery).)* ## Brief change log *(for example:)* - *The TaskInfo is stored in the blob store on job creation time as a persistent artifact* - *Deployments RPC transmits only the blob storage reference* - *TaskManagers retrieve the TaskInfo from the blob cache* ## Verifying this change Please make sure both new and modified tests in this PR follows the conventions defined in our code quality guide: https://flink.apache.org/contributing/code-style-and-quality-common.html#testing *(Please pick either of the following options)* This change is a trivial rework / code cleanup without any test coverage. *(or)* This change is already covered by existing tests, such as *(please describe tests)*. *(or)* This change added tests and can be verified as follows: *(example:)* - *Added integration tests for end-to-end deployment with large payloads (100MB)* - *Extended integration test for recovery after master (JobManager) failure* - *Added test that validates that TaskInfo is transferred only once across recoveries* - *Manually verified the change by running a 4 node cluser with 2 JobManagers and 4 TaskManagers, a stateful streaming program, and killing one JobManager and two TaskManagers during the execution, verifying that recovery happens correctly.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / no) - The serializers: (yes / no / don't know) - The runtime per-record code paths (performance sensitive): (yes / no / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know) - The S3 file system connector: (yes / no / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / no) - If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18336: [FLINK-25441][network] Wrap failure cuase with ProducerFailedException only for PipelinedSubpartitionView
flinkbot edited a comment on pull request #18336: URL: https://github.com/apache/flink/pull/18336#issuecomment-1010777588 ## CI report: * 86b6155caba9160268797213db5e9c0a18c95a09 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29312) * 1a960b7f82e85d42e2a4f5eb9e9d63a203ed9de5 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29358) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #18336: [FLINK-25441][network] Wrap failure cuase with ProducerFailedException only for PipelinedSubpartitionView
flinkbot edited a comment on pull request #18336: URL: https://github.com/apache/flink/pull/18336#issuecomment-1010777588 ## CI report: * 86b6155caba9160268797213db5e9c0a18c95a09 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29312) * 1a960b7f82e85d42e2a4f5eb9e9d63a203ed9de5 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wanglijie95 commented on pull request #18336: [FLINK-25441][network] Wrap failure cuase with ProducerFailedException only for PipelinedSubpartitionView
wanglijie95 commented on pull request #18336: URL: https://github.com/apache/flink/pull/18336#issuecomment-1011797394 > Changes LGTM. Though the test failure seems unrelated, could please you re-trigger it and get a green build? @wanglijie95 W Sure, and I will squash the commits. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry commented on pull request #18336: [FLINK-25441][network] Wrap failure cuase with ProducerFailedException only for PipelinedSubpartitionView
wsry commented on pull request #18336: URL: https://github.com/apache/flink/pull/18336#issuecomment-1011790899 Changes LGTM. Though the test failure seems unrelated, could please you re-trigger it and get a green build? @wanglijie95 W -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot edited a comment on pull request #17676: [FLINK-14100][connectors] Add Oracle dialect
flinkbot edited a comment on pull request #17676: URL: https://github.com/apache/flink/pull/17676#issuecomment-960641444 ## CI report: * 715383d5d8ec50ae6fd41db834d800bb65b99c80 Azure: [FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29095) * 35238e2baa5a30a5a5e047aceddedbfae6cfcd75 Azure: [PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=29357) Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org