[jira] [Reopened] (FLINK-21207) 'csv.disable-quote-character' can not take effect during deserialization

2021-07-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21207?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee reopened FLINK-21207:
--

> 'csv.disable-quote-character'  can not take effect during deserialization
> -
>
> Key: FLINK-21207
> URL: https://issues.apache.org/jira/browse/FLINK-21207
> Project: Flink
>  Issue Type: New Feature
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Ecosystem
>Reporter: sharkd tu
>Assignee: sharkd tu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> source table with 'csv.disable-quote-character' = 'true' can not take effect.
>  
> {code}
> Caused by: 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.JsonParseException:
>  Missing closing quote for valueCaused by: 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.JsonParseException:
>  Missing closing quote for value at [Source: UNKNOWN; line: 1, column: 29] at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.JsonParser._constructError(JsonParser.java:1840)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.base.ParserMinimalBase._reportError(ParserMinimalBase.java:712)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvParser._reportParsingError(CsvParser.java:1250)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.impl.CsvDecoder._nextQuotedString(CsvDecoder.java:785)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.impl.CsvDecoder.nextString(CsvDecoder.java:630)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvParser._handleNextEntry(CsvParser.java:846)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvParser.nextFieldName(CsvParser.java:665)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.std.BaseNodeDeserializer.deserializeObject(JsonNodeDeserializer.java:249)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.std.JsonNodeDeserializer.deserialize(JsonNodeDeserializer.java:68)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.std.JsonNodeDeserializer.deserialize(JsonNodeDeserializer.java:15)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader._bindAndClose(ObjectReader.java:1719)
>  at 
> org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader.readValue(ObjectReader.java:1282)
>  at 
> org.apache.flink.formats.csv.CsvRowDataDeserializationSchema.deserialize(CsvRowDataDeserializationSchema.java:155)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-23490) Flink Table Example - StreamWindowSQLExample shows output in older format

2021-07-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23490?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-23490.

Resolution: Fixed

master: a23d8e393a9179f90ce879d8ba21ee36a946360e

> Flink Table Example - StreamWindowSQLExample shows output in older format
> -
>
> Key: FLINK-23490
> URL: https://issues.apache.org/jira/browse/FLINK-23490
> Project: Flink
>  Issue Type: Improvement
>  Components: Examples, Table SQL / API
>Affects Versions: 1.13.1
>Reporter: Mans Singh
>Assignee: Mans Singh
>Priority: Minor
>  Labels: examples, pull-request-available, sql, table
> Fix For: 1.14.0
>
>   Original Estimate: 0.25h
>  Remaining Estimate: 0.25h
>
> The example print output shows older format:
> {quote}{{// 2019-12-12 00:00:00.000,3,10,3}}
>  {{// 2019-12-12 00:00:05.000,3,6,2}}
> {quote}
>  
>  Execution of the application print the following:
> {quote}+I[2019-12-12 00:00:00.000, 3, 10, 3]
>  +I[2019-12-12 00:00:05.000, 3, 6, 2]
> {quote}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi merged pull request #16587: [FLINK-23490][examples][table] - Updated print output

2021-07-27 Thread GitBox


JingsongLi merged pull request #16587:
URL: https://github.com/apache/flink/pull/16587


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] JingsongLi merged pull request #16462: [FLINK-22985][table-runtime] Fix NullPointerException when comparing temporal type with invalid string literal

2021-07-27 Thread GitBox


JingsongLi merged pull request #16462:
URL: https://github.com/apache/flink/pull/16462


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Thesharing commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


Thesharing commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677984188



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobCacheSizeTracker.java
##
@@ -99,44 +101,56 @@ public BlobCacheSizeTracker(long sizeLimit) {
 }
 
 /** Register the target file to the tracker. */
-public void track(@Nullable JobID jobId, BlobKey blobKey, long size) {
+public void track(JobID jobId, BlobKey blobKey, long size) {
+checkNotNull(jobId);
 checkNotNull(blobKey);
 checkArgument(size >= 0);
 
 synchronized (lock) {
-caches.put(Tuple2.of(jobId, blobKey), size);
-if (jobId != null) {
+if (caches.putIfAbsent(Tuple2.of(jobId, blobKey), size) == null) {
 blobKeyByJob.computeIfAbsent(jobId, ignore -> new 
HashSet<>()).add(blobKey);
-}
-total += size;
-if (total > sizeLimit) {
+
+total += size;
+if (total > sizeLimit) {
+LOG.warn(
+"The overall size of ShuffleDescriptors in 
PermanentBlobCache exceeds "
++ "the limit. Limit = [{}], Current: [{}], 
"
++ "The size of next ShuffleDescriptors: 
[{}].",
+sizeLimit,
+total,
+size);
+}
+} else {
 LOG.warn(

Review comment:
   I think this warning shouldn't happen frequently. In 
`BlobUtil.moveTempFileToStore`, if the blob key duplicates, a warning will be 
outputted, too. Furthermore, the previous warning ("The overall size of 
ShuffleDescriptors in PermanentBlobCache exceeds ...") should be warning, too. 
The overall size exceeds if and only if the size of blob exceeds 100MiB. In 
this scenario the user should be warned about it.

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/PermanentBlobService.java
##
@@ -50,13 +50,20 @@
 /**
  * Returns the content of the file for the BLOB with the provided job ID 
the blob key.
  *
+ * Compared to {@code getFile}, {@code readFile} will attempt to read 
the entire file after
+ * retrieving it. If file reading and file retrieving is done in the same 
WRITE lock, it can
+ * avoid the scenario that the path to the file is deleted concurrently by 
other threads at the
+ * same time as the file is retrieved and read.

Review comment:
   Thank you for pointing this out. Resolved.

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobCacheSizeTracker.java
##
@@ -99,44 +101,56 @@ public BlobCacheSizeTracker(long sizeLimit) {
 }
 
 /** Register the target file to the tracker. */
-public void track(@Nullable JobID jobId, BlobKey blobKey, long size) {
+public void track(JobID jobId, BlobKey blobKey, long size) {
+checkNotNull(jobId);
 checkNotNull(blobKey);
 checkArgument(size >= 0);
 
 synchronized (lock) {
-caches.put(Tuple2.of(jobId, blobKey), size);
-if (jobId != null) {
+if (caches.putIfAbsent(Tuple2.of(jobId, blobKey), size) == null) {
 blobKeyByJob.computeIfAbsent(jobId, ignore -> new 
HashSet<>()).add(blobKey);
-}
-total += size;
-if (total > sizeLimit) {
+
+total += size;
+if (total > sizeLimit) {
+LOG.warn(
+"The overall size of ShuffleDescriptors in 
PermanentBlobCache exceeds "
++ "the limit. Limit = [{}], Current: [{}], 
"
++ "The size of next ShuffleDescriptors: 
[{}].",

Review comment:
   Resolved.

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobCacheSizeTracker.java
##
@@ -99,44 +101,56 @@ public BlobCacheSizeTracker(long sizeLimit) {
 }
 
 /** Register the target file to the tracker. */
-public void track(@Nullable JobID jobId, BlobKey blobKey, long size) {
+public void track(JobID jobId, BlobKey blobKey, long size) {
+checkNotNull(jobId);
 checkNotNull(blobKey);
 checkArgument(size >= 0);
 
 synchronized (lock) {
-caches.put(Tuple2.of(jobId, blobKey), size);
-if (jobId != null) {
+if (caches.putIfAbsent(Tuple2.of(jobId, blobKey), size) == null) {
 blobKeyByJob.computeIfAbsent(jobId, ignore -> new 
HashSet<>()).add(blobKey);
-}
-total += size;
-if (total > sizeLimit) {
+
+total += size;
+if (total > sizeLimit) {
+LOG.warn(
+"The overall size of ShuffleDescriptors in 

[GitHub] [flink] huxixiang commented on pull request #16547: [FLINK-23423][docs-zh] Translate the page of "Elasticsearch Connector" into Chinese

2021-07-27 Thread GitBox


huxixiang commented on pull request #16547:
URL: https://github.com/apache/flink/pull/16547#issuecomment-888022807


   cc @95chenjz ,Hi jianzhang,could you review my PR when you're free? thanks a 
lot.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16614: [FLINK-21207][csv] Fix 'csv.disable-quote-character' = 'true' can not…

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16614:
URL: https://github.com/apache/flink/pull/16614#issuecomment-887991835


   
   ## CI report:
   
   * 3e187d233039c48dc9457697bf2a60253789c9f6 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21066)
 
   * bf0fe365b3f9e3f10c49803b2e0c089171cf7de1 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Thesharing commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


Thesharing commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677982428



##
File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/DefaultExecutionGraphDeploymentWithSizeLimitTest.java
##
@@ -0,0 +1,185 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License
+ */
+
+package org.apache.flink.runtime.executiongraph;
+
+import org.apache.flink.configuration.BlobServerOptions;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.runtime.JobException;
+import org.apache.flink.runtime.blob.BlobCacheSizeTracker;
+import org.apache.flink.runtime.blob.BlobServer;
+import org.apache.flink.runtime.blob.PermanentBlobCache;
+import org.apache.flink.runtime.blob.VoidBlobStore;
+import org.apache.flink.runtime.client.JobExecutionException;
+import 
org.apache.flink.runtime.concurrent.ComponentMainThreadExecutorServiceAdapter;
+import org.apache.flink.runtime.deployment.InputGateDeploymentDescriptor;
+import org.apache.flink.runtime.deployment.TaskDeploymentDescriptor;
+import org.apache.flink.runtime.execution.ExecutionState;
+import 
org.apache.flink.runtime.executiongraph.utils.SimpleAckingTaskManagerGateway;
+import org.apache.flink.runtime.io.network.partition.ResultPartitionType;
+import org.apache.flink.runtime.jobgraph.DistributionPattern;
+import org.apache.flink.runtime.jobgraph.IntermediateResultPartitionID;
+import org.apache.flink.runtime.jobgraph.JobGraph;
+import org.apache.flink.runtime.jobgraph.JobGraphTestUtils;
+import org.apache.flink.runtime.jobgraph.JobVertex;
+import org.apache.flink.runtime.jobgraph.JobVertexID;
+import org.apache.flink.runtime.jobmaster.LogicalSlot;
+import org.apache.flink.runtime.jobmaster.TestingLogicalSlotBuilder;
+import org.apache.flink.runtime.operators.BatchTask;
+import org.apache.flink.runtime.scheduler.strategy.ConsumedPartitionGroup;
+import org.apache.flink.runtime.testutils.DirectScheduledExecutorService;
+import org.apache.flink.util.function.FunctionUtils;
+
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
+
+import java.io.IOException;
+import java.net.InetSocketAddress;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.ArrayBlockingQueue;
+import java.util.concurrent.BlockingQueue;
+
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNotNull;
+
+public class DefaultExecutionGraphDeploymentWithSizeLimitTest
+extends DefaultExecutionGraphDeploymentWithBlobCacheTest {
+
+@Before
+@Override
+public void setupBlobServer() throws IOException {
+Configuration config = new Configuration();
+// Always offload the serialized JobInformation, TaskInformation and 
cached
+// ShuffleDescriptors
+config.setInteger(BlobServerOptions.OFFLOAD_MINSIZE, 0);
+blobServer = new BlobServer(config, new VoidBlobStore());
+blobServer.start();
+blobWriter = blobServer;
+
+InetSocketAddress serverAddress = new InetSocketAddress("localhost", 
blobServer.getPort());
+// Set the size limit of the blob cache to 1
+BlobCacheSizeTracker blobCacheSizeTracker = new 
BlobCacheSizeTracker(1L);
+blobCache =
+new PermanentBlobCache(
+config, new VoidBlobStore(), serverAddress, 
blobCacheSizeTracker);
+}
+
+@After
+@Override
+public void shutdownBlobServer() throws IOException {
+if (blobServer != null) {
+blobServer.close();
+}
+}
+
+@Test
+public void testDeployTasksWithMinimumSizeLimit() throws Exception {

Review comment:
   Sure. Resolved.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15322: [FLINK-21353][state] Add DFS-based StateChangelog (TM-owned state)

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #15322:
URL: https://github.com/apache/flink/pull/15322#issuecomment-804015738


   
   ## CI report:
   
   * eb50427eca80dd62004e23945d5bbd7fadb97319 UNKNOWN
   * 25c523343780cd96be23cb394830a91cc1fb200b Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=20580)
 
   * 16663a4ee0f5ff545903d5669a5d640f2953e109 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Thesharing commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


Thesharing commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677979942



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/PermanentBlobCache.java
##
@@ -187,6 +211,139 @@ public File getFile(JobID jobId, PermanentBlobKey key) 
throws IOException {
 return getFileInternal(jobId, key);
 }
 
+/**
+ * Returns the content of the file for the BLOB with the provided job ID 
the blob key.
+ *
+ * The method will first attempt to serve the BLOB from the local 
cache. If the BLOB is not
+ * in the cache, the method will try to download it from the HA store, or 
directly from the
+ * {@link BlobServer}.
+ *
+ * Compared to {@code getFile}, {@code readFile} makes sure that the 
file is fully read in
+ * the same write lock as the file is accessed. This avoids the scenario 
that the path is
+ * returned as the file is deleted concurrently by other threads.
+ *
+ * @param jobId ID of the job this blob belongs to
+ * @param blobKey BLOB key associated with the requested file
+ * @return The content of the BLOB.
+ * @throws java.io.FileNotFoundException if the BLOB does not exist;
+ * @throws IOException if any other error occurs when retrieving the file.
+ */
+@Override
+public byte[] readFile(JobID jobId, PermanentBlobKey blobKey) throws 
IOException {
+checkNotNull(jobId);
+checkNotNull(blobKey);
+
+final File localFile = BlobUtils.getStorageLocation(storageDir, jobId, 
blobKey);
+readWriteLock.readLock().lock();
+
+try {
+if (localFile.exists()) {
+blobCacheSizeTracker.update(jobId, blobKey);
+return FileUtils.readAllBytes(localFile.toPath());
+}
+} finally {
+readWriteLock.readLock().unlock();
+}
+
+// first try the distributed blob store (if available)
+// use a temporary file (thread-safe without locking)
+File incomingFile = createTemporaryFilename();
+try {
+try {
+if (blobView.get(jobId, blobKey, incomingFile)) {
+// now move the temp file to our local cache atomically
+readWriteLock.writeLock().lock();
+try {
+checkLimitAndMoveFile(incomingFile, jobId, blobKey, 
localFile, log, null);
+return FileUtils.readAllBytes(localFile.toPath());
+} finally {
+readWriteLock.writeLock().unlock();
+}
+}
+} catch (Exception e) {
+log.info(
+"Failed to copy from blob store. Downloading from BLOB 
server instead.", e);
+}
+
+final InetSocketAddress currentServerAddress = serverAddress;
+
+if (currentServerAddress != null) {
+// fallback: download from the BlobServer
+BlobClient.downloadFromBlobServer(
+jobId,
+blobKey,
+incomingFile,
+currentServerAddress,
+blobClientConfig,
+numFetchRetries);
+
+readWriteLock.writeLock().lock();
+try {
+checkLimitAndMoveFile(incomingFile, jobId, blobKey, 
localFile, log, null);
+return FileUtils.readAllBytes(localFile.toPath());
+} finally {
+readWriteLock.writeLock().unlock();
+}
+} else {
+throw new IOException(
+"Cannot download from BlobServer, because the server 
address is unknown.");
+}
+
+} finally {
+// delete incomingFile from a failed download
+if (!incomingFile.delete() && incomingFile.exists()) {
+log.warn(
+"Could not delete the staging file {} for blob key {} 
and job {}.",
+incomingFile,
+blobKey,
+jobId);
+}
+}
+}
+
+private void checkLimitAndMoveFile(
+File incomingFile,
+@Nullable JobID jobId,
+BlobKey blobKey,
+File localFile,
+Logger log,
+@Nullable BlobStore blobStore)
+throws IOException {
+
+// Check the size limit and delete the files that exceeds the limit
+final long sizeOfIncomingFile = incomingFile.length();
+final List> blobsToDelete =
+blobCacheSizeTracker.checkLimit(sizeOfIncomingFile);
+
+for (Tuple2 key : blobsToDelete) {
+deleteFile(key.f0, key.f1);
+blobCacheSizeTracker.untrack(key);

Review comment:
   I'm not sure the meaning of exposing 

[GitHub] [flink] flinkbot edited a comment on pull request #16587: [FLINK-23490][examples][table] - Updated print output

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16587:
URL: https://github.com/apache/flink/pull/16587#issuecomment-886122043


   
   ## CI report:
   
   * 3eedc57632b94d0832cf308b58fa4c2008462f6b UNKNOWN
   * f38ef3af364109bffbdbd5cc8b812c40b2af3eb9 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21057)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rkhachatryan commented on pull request #15322: [FLINK-21353][state] Add DFS-based StateChangelog (TM-owned state)

2021-07-27 Thread GitBox


rkhachatryan commented on pull request #15322:
URL: https://github.com/apache/flink/pull/15322#issuecomment-888008814


   Great, then I'll merge the PR once the CI completes.
   (I've squashed the commits and rebased onto the latest master).


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] tweise commented on a change in pull request #15924: [FLINK-22670][FLIP-150][connector/common] Hybrid source baseline

2021-07-27 Thread GitBox


tweise commented on a change in pull request #15924:
URL: https://github.com/apache/flink/pull/15924#discussion_r677967467



##
File path: 
flink-connectors/flink-connector-base/src/main/java/org/apache/flink/connector/base/source/hybrid/HybridSource.java
##
@@ -110,69 +106,64 @@ public Boundedness getBoundedness() {
 
 @Override
 public SimpleVersionedSerializer getSplitSerializer() {
-// List> serializers = new 
ArrayList<>();
-// TODO: serializers are created on demand as underlying sources are 
created during switch
-// sources.forEach(t -> 
serializers.add(castSerializer(t.source.getSplitSerializer(;
 return new HybridSourceSplitSerializer(switchedSources);
 }
 
 @Override
 public SimpleVersionedSerializer
 getEnumeratorCheckpointSerializer() {
-List> serializers = new 
ArrayList<>();
-sources.forEach(
-t -> 
serializers.add(castSerializer(t.source.getEnumeratorCheckpointSerializer(;
-return new HybridSourceEnumeratorStateSerializer(serializers);
-}
-
-private static  SimpleVersionedSerializer castSerializer(
-SimpleVersionedSerializer s) {
-@SuppressWarnings("rawtypes")
-SimpleVersionedSerializer s1 = s;
-return s1;
+return new HybridSourceEnumeratorStateSerializer(switchedSources);
 }
 
 /**
- * Callback for switch time customization of the underlying source, 
typically to dynamically set
- * a start position from previous enumerator end state.
+ * Factory for underlying sources of {@link HybridSource}.
  *
- * Requires the ability to augment the existing source (or clone and 
modify). Provides the
- * flexibility to set start position in any way a source allows, in a 
source specific way.
- * Future convenience could be built on top of it, for example an 
implementation recognizes
- * optional interfaces.
+ * This factory permits building of a source at graph construction time 
or deferred at switch
+ * time. Provides the ability to set a start position in any way a 
specific source allows.
+ * Future convenience could be built on top of it, for example a default 
implementation that
+ * recognizes optional interfaces to transfer position in a common format.
  *
  * Called when the current enumerator has finished and before the next 
enumerator is created.
- * The enumerator end state can thus be used to set the next source's 
start start position.
+ * The enumerator end state can thus be used to set the next source's 
start start position. Only
+ * required for dynamic position transfer at time of switching.
  *
- * Only required for dynamic position transfer at time of switching, 
otherwise source can be
- * preconfigured with a start position during job submission.
+ * If start position is known at jib submission, the source can be 
constructed in the entry
+ * point and simply wrapped into the factory, providing the benefit of 
validation during
+ * submission.
  */
-public interface SourceConfigurer
+public interface SourceFactory
 extends Serializable {
-SourceT configure(SourceT source, FromEnumT enumerator);
+SourceT create(FromEnumT enumerator);

Review comment:
   @becketqin your description of the switching process is correct but it 
only covers the scenario where the start position of the subsequent source is 
determined at switch time. There is the much simpler scenario with fixed start 
position which does not require the user to implement `SourceFactory`:
   
   ```
   FileSource fileSource =
   FileSource.forRecordStreamFormat(new TextLineFormat(), 
Path.fromLocalFile(testDir)).build();
   KafkaSource kafkaSource =
   KafkaSource.builder()
   .setBootstrapServers("localhost:9092")
   .setGroupId("MyGroup")
   .setTopics(Arrays.asList("quickstart-events"))
   .setDeserializer(
   
KafkaRecordDeserializer.valueOnly(StringDeserializer.class))
   .setStartingOffsets(OffsetsInitializer.earliest())
   .build();
   HybridSource hybridSource =
   HybridSource.builder(fileSource)
   .addSource(kafkaSource)
   .build();
   
   ```
   
   The intention for this PR is to cover that simple scenario and at the same 
time provide the mechanism to build something more advanced with a different 
bounded source. This isn't in conflict with subsequent addition of standard 
interface for extracting end position and we can discuss whether or not that 
should be part of FLIP-150. However I hope to find a way to move this PR 
forward independently as it has been under review for a long time 

[GitHub] [flink] flinkbot edited a comment on pull request #16614: [FLINK-21207][csv] Fix 'csv.disable-quote-character' = 'true' can not…

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16614:
URL: https://github.com/apache/flink/pull/16614#issuecomment-887991835


   
   ## CI report:
   
   * 3e187d233039c48dc9457697bf2a60253789c9f6 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21066)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16552: [FLINK-23449] [Deployment/YARN] YarnTaskExecutorRunner should contains MapReduce classes

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16552:
URL: https://github.com/apache/flink/pull/16552#issuecomment-884119711


   
   ## CI report:
   
   * 59856a6793c6b6956879899d2a61554dc7a3 UNKNOWN
   * 3f6d7224e9cb4d78e17354f0d6830286e5f1c069 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=20816)
 
   * f37155567bd0accbd44972db3650eae9c5d893cf Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21065)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #16614: [FLINK-21207][csv] Fix 'csv.disable-quote-character' = 'true' can not…

2021-07-27 Thread GitBox


flinkbot commented on pull request #16614:
URL: https://github.com/apache/flink/pull/16614#issuecomment-887991835


   
   ## CI report:
   
   * 3e187d233039c48dc9457697bf2a60253789c9f6 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16588: [FLINK-22936] Support column comment in Schema and ResolvedSchema

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16588:
URL: https://github.com/apache/flink/pull/16588#issuecomment-886186049


   
   ## CI report:
   
   * 87b1a1cdd26823c9db6069c40a8c842059012311 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21032)
 
   * 28774cb45a375dde25c18e73bbfd7afa684ee151 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21064)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16552: [FLINK-23449] [Deployment/YARN] YarnTaskExecutorRunner should contains MapReduce classes

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16552:
URL: https://github.com/apache/flink/pull/16552#issuecomment-884119711


   
   ## CI report:
   
   * 59856a6793c6b6956879899d2a61554dc7a3 UNKNOWN
   * 3f6d7224e9cb4d78e17354f0d6830286e5f1c069 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=20816)
 
   * f37155567bd0accbd44972db3650eae9c5d893cf UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-23508) HA mode cannot coexist with history server

2021-07-27 Thread Yang Wang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Wang closed FLINK-23508.
-
Resolution: Duplicate

This is a valid issue and see more discussion in FLINK-19358.

> HA mode cannot coexist with history server
> --
>
> Key: FLINK-23508
> URL: https://issues.apache.org/jira/browse/FLINK-23508
> Project: Flink
>  Issue Type: Bug
>Reporter: john
>Priority: Major
>
> When I submit a Flink job using HA mode, the default job ID generated each 
> time is . At this time, I cannot use the 
> history server because the history server uses the job ID to write the job 
> archive. But the job ID is the same every time, so I cannot track the status 
> of the job. what do I do?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-statefun] sjwiesman commented on pull request #244: [FLINK-23520][datastream] Improve StateFun <-> DataStream Interop

2021-07-27 Thread GitBox


sjwiesman commented on pull request #244:
URL: https://github.com/apache/flink-statefun/pull/244#issuecomment-887988944


   Closing this because I realized I can simplify the type conversion and I 
don’t want to wate anyone’s time reviewing something I change tomorrow :)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] sjwiesman closed pull request #244: [FLINK-23520][datastream] Improve StateFun <-> DataStream Interop

2021-07-27 Thread GitBox


sjwiesman closed pull request #244:
URL: https://github.com/apache/flink-statefun/pull/244


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #16614: [FLINK-21207][csv] Fix 'csv.disable-quote-character' = 'true' can not…

2021-07-27 Thread GitBox


flinkbot commented on pull request #16614:
URL: https://github.com/apache/flink/pull/16614#issuecomment-887984015


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 3e187d233039c48dc9457697bf2a60253789c9f6 (Wed Jul 28 
03:36:58 UTC 2021)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] chenlei677 opened a new pull request #16614: [FLINK-21207][csv] Fix 'csv.disable-quote-character' = 'true' can not…

2021-07-27 Thread GitBox


chenlei677 opened a new pull request #16614:
URL: https://github.com/apache/flink/pull/16614


   … take effect in source table
   
   This closes #14813
   
   
   
   ## What is the purpose of the change
   Fix source table with 'csv.disable-quote-character' = 'true' can not take 
effect in flink 1.12.
   
   `Caused by: 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.JsonParseException: 
Unexpected character ('n' (code 110)): Expected separator ('"' (code 34)) or 
end-of-line
at [Source: 
(org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.impl.UTF8Reader);
 line: 1, column: 6]
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.JsonParser._constructError(JsonParser.java:1840)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.base.ParserMinimalBase._reportError(ParserMinimalBase.java:712)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.core.base.ParserMinimalBase._reportUnexpectedChar(ParserMinimalBase.java:637)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvParser._reportUnexpectedCsvChar(CsvParser.java:1254)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.impl.CsvDecoder._nextQuotedString(CsvDecoder.java:867)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.impl.CsvDecoder.nextString(CsvDecoder.java:630)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvParser._handleNextEntry(CsvParser.java:846)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.dataformat.csv.CsvParser.nextFieldName(CsvParser.java:665)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.std.BaseNodeDeserializer.deserializeObject(JsonNodeDeserializer.java:249)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.std.JsonNodeDeserializer.deserialize(JsonNodeDeserializer.java:68)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.deser.std.JsonNodeDeserializer.deserialize(JsonNodeDeserializer.java:15)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader._bindAndClose(ObjectReader.java:1719)
at 
org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.ObjectReader.readValue(ObjectReader.java:1282)
at 
org.apache.flink.formats.csv.CsvRowDataDeserializationSchema.deserialize(CsvRowDataDeserializationSchema.java:145)
... 9 more`
   
   ## Brief change log
   
   Invoke disableQuoteChar of CsvSchema when 'disable-quote-character' is 
enabled.
   
   ## Verifying this change
   
   It can be verified in 
CsvFormatFactoryTest#testDeserializeWithDisableQuoteCharacter.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


zhuzhurk commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677945612



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/PermanentBlobCache.java
##
@@ -247,6 +404,7 @@ public void run() {
 
 boolean success = false;
 try {
+blobCacheSizeTracker.unregisterJob(jobId);

Review comment:
   Thanks for the explanation. I think you are right.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


zhuzhurk commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677943607



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobCacheSizeTracker.java
##
@@ -99,44 +101,56 @@ public BlobCacheSizeTracker(long sizeLimit) {
 }
 
 /** Register the target file to the tracker. */
-public void track(@Nullable JobID jobId, BlobKey blobKey, long size) {
+public void track(JobID jobId, BlobKey blobKey, long size) {
+checkNotNull(jobId);
 checkNotNull(blobKey);
 checkArgument(size >= 0);
 
 synchronized (lock) {
-caches.put(Tuple2.of(jobId, blobKey), size);
-if (jobId != null) {
+if (caches.putIfAbsent(Tuple2.of(jobId, blobKey), size) == null) {
 blobKeyByJob.computeIfAbsent(jobId, ignore -> new 
HashSet<>()).add(blobKey);
-}
-total += size;
-if (total > sizeLimit) {
+
+total += size;
+if (total > sizeLimit) {
+LOG.warn(
+"The overall size of ShuffleDescriptors in 
PermanentBlobCache exceeds "

Review comment:
   `ShuffleDescriptors` -> `blob`

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobCacheSizeTracker.java
##
@@ -99,44 +101,56 @@ public BlobCacheSizeTracker(long sizeLimit) {
 }
 
 /** Register the target file to the tracker. */
-public void track(@Nullable JobID jobId, BlobKey blobKey, long size) {
+public void track(JobID jobId, BlobKey blobKey, long size) {
+checkNotNull(jobId);
 checkNotNull(blobKey);
 checkArgument(size >= 0);
 
 synchronized (lock) {
-caches.put(Tuple2.of(jobId, blobKey), size);
-if (jobId != null) {
+if (caches.putIfAbsent(Tuple2.of(jobId, blobKey), size) == null) {
 blobKeyByJob.computeIfAbsent(jobId, ignore -> new 
HashSet<>()).add(blobKey);
-}
-total += size;
-if (total > sizeLimit) {
+
+total += size;
+if (total > sizeLimit) {
+LOG.warn(
+"The overall size of ShuffleDescriptors in 
PermanentBlobCache exceeds "
++ "the limit. Limit = [{}], Current: [{}], 
"
++ "The size of next ShuffleDescriptors: 
[{}].",

Review comment:
   `ShuffleDescriptors` -> `blob`

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobCacheSizeTracker.java
##
@@ -99,44 +101,56 @@ public BlobCacheSizeTracker(long sizeLimit) {
 }
 
 /** Register the target file to the tracker. */
-public void track(@Nullable JobID jobId, BlobKey blobKey, long size) {
+public void track(JobID jobId, BlobKey blobKey, long size) {
+checkNotNull(jobId);
 checkNotNull(blobKey);
 checkArgument(size >= 0);
 
 synchronized (lock) {
-caches.put(Tuple2.of(jobId, blobKey), size);
-if (jobId != null) {
+if (caches.putIfAbsent(Tuple2.of(jobId, blobKey), size) == null) {
 blobKeyByJob.computeIfAbsent(jobId, ignore -> new 
HashSet<>()).add(blobKey);
-}
-total += size;
-if (total > sizeLimit) {
+
+total += size;
+if (total > sizeLimit) {
+LOG.warn(
+"The overall size of ShuffleDescriptors in 
PermanentBlobCache exceeds "
++ "the limit. Limit = [{}], Current: [{}], 
"
++ "The size of next ShuffleDescriptors: 
[{}].",
+sizeLimit,
+total,
+size);
+}
+} else {
 LOG.warn(

Review comment:
   Is it possible to happen frequently. If yes, I prefer to make it a debug 
log.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16613: [BP-1.13][FLINK-23418][e2e] Increase the timeout to make kubernetes application ha test more stable

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16613:
URL: https://github.com/apache/flink/pull/16613#issuecomment-887972126


   
   ## CI report:
   
   * 51a7fc3d1abada9a68e4249da55853c5fa1cc1e2 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21063)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16588: [FLINK-22936] Support column comment in Schema and ResolvedSchema

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16588:
URL: https://github.com/apache/flink/pull/16588#issuecomment-886186049


   
   ## CI report:
   
   * 87b1a1cdd26823c9db6069c40a8c842059012311 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21032)
 
   * 28774cb45a375dde25c18e73bbfd7afa684ee151 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


zhuzhurk commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677941649



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/PermanentBlobCache.java
##
@@ -187,6 +211,139 @@ public File getFile(JobID jobId, PermanentBlobKey key) 
throws IOException {
 return getFileInternal(jobId, key);
 }
 
+/**
+ * Returns the content of the file for the BLOB with the provided job ID 
the blob key.
+ *
+ * The method will first attempt to serve the BLOB from the local 
cache. If the BLOB is not
+ * in the cache, the method will try to download it from the HA store, or 
directly from the
+ * {@link BlobServer}.
+ *
+ * Compared to {@code getFile}, {@code readFile} makes sure that the 
file is fully read in
+ * the same write lock as the file is accessed. This avoids the scenario 
that the path is
+ * returned as the file is deleted concurrently by other threads.
+ *
+ * @param jobId ID of the job this blob belongs to
+ * @param blobKey BLOB key associated with the requested file
+ * @return The content of the BLOB.
+ * @throws java.io.FileNotFoundException if the BLOB does not exist;
+ * @throws IOException if any other error occurs when retrieving the file.
+ */
+@Override
+public byte[] readFile(JobID jobId, PermanentBlobKey blobKey) throws 
IOException {
+checkNotNull(jobId);
+checkNotNull(blobKey);
+
+final File localFile = BlobUtils.getStorageLocation(storageDir, jobId, 
blobKey);
+readWriteLock.readLock().lock();
+
+try {
+if (localFile.exists()) {
+blobCacheSizeTracker.update(jobId, blobKey);
+return FileUtils.readAllBytes(localFile.toPath());
+}
+} finally {
+readWriteLock.readLock().unlock();
+}
+
+// first try the distributed blob store (if available)
+// use a temporary file (thread-safe without locking)
+File incomingFile = createTemporaryFilename();
+try {
+try {
+if (blobView.get(jobId, blobKey, incomingFile)) {
+// now move the temp file to our local cache atomically
+readWriteLock.writeLock().lock();
+try {
+checkLimitAndMoveFile(incomingFile, jobId, blobKey, 
localFile, log, null);
+return FileUtils.readAllBytes(localFile.toPath());
+} finally {
+readWriteLock.writeLock().unlock();
+}
+}
+} catch (Exception e) {
+log.info(
+"Failed to copy from blob store. Downloading from BLOB 
server instead.", e);
+}
+
+final InetSocketAddress currentServerAddress = serverAddress;
+
+if (currentServerAddress != null) {
+// fallback: download from the BlobServer
+BlobClient.downloadFromBlobServer(
+jobId,
+blobKey,
+incomingFile,
+currentServerAddress,
+blobClientConfig,
+numFetchRetries);
+
+readWriteLock.writeLock().lock();
+try {
+checkLimitAndMoveFile(incomingFile, jobId, blobKey, 
localFile, log, null);
+return FileUtils.readAllBytes(localFile.toPath());
+} finally {
+readWriteLock.writeLock().unlock();
+}
+} else {
+throw new IOException(
+"Cannot download from BlobServer, because the server 
address is unknown.");
+}
+
+} finally {
+// delete incomingFile from a failed download
+if (!incomingFile.delete() && incomingFile.exists()) {
+log.warn(
+"Could not delete the staging file {} for blob key {} 
and job {}.",
+incomingFile,
+blobKey,
+jobId);
+}
+}
+}
+
+private void checkLimitAndMoveFile(
+File incomingFile,
+@Nullable JobID jobId,

Review comment:
   We should also add `checkNotNull` for it.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] curcur commented on a change in pull request #16531: [FLINK-23204] Provide StateBackends access to MailboxExecutor

2021-07-27 Thread GitBox


curcur commented on a change in pull request #16531:
URL: https://github.com/apache/flink/pull/16531#discussion_r677941509



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/StreamTask.java
##
@@ -338,6 +338,9 @@ protected StreamTask(
 Executors.newCachedThreadPool(
 new ExecutorThreadFactory("AsyncOperations", 
uncaughtExceptionHandler));
 
+environment.setMainMailboxExecutor(mainMailboxExecutor);
+environment.setAsyncOperationsThreadPool(asyncOperationsThreadPool);

Review comment:
   > I think it's not exactly true - asyncOperationsThreadPool only waits 
for snapshot completion, but upload can happen in other threads (which is the 
with RocksDB).
   
   That's exactly happening here: `asyncOperationsThreadPool` is waiting for a 
snapshot completion. The difference is that for the normal state backend, async 
happens during checkpointing; while in materialization(changelog), it happens 
during periodic materialization.
   
   > Merging these pools is out of scope (if it happens), but should be taken 
into account IMO.
   That part I totally agree, that why `asyncOperationsThreadPool` is reused 
here.
   




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


zhuzhurk commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677940469



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/PermanentBlobCache.java
##
@@ -187,6 +211,139 @@ public File getFile(JobID jobId, PermanentBlobKey key) 
throws IOException {
 return getFileInternal(jobId, key);
 }
 
+/**
+ * Returns the content of the file for the BLOB with the provided job ID 
the blob key.
+ *
+ * The method will first attempt to serve the BLOB from the local 
cache. If the BLOB is not
+ * in the cache, the method will try to download it from the HA store, or 
directly from the
+ * {@link BlobServer}.
+ *
+ * Compared to {@code getFile}, {@code readFile} makes sure that the 
file is fully read in
+ * the same write lock as the file is accessed. This avoids the scenario 
that the path is
+ * returned as the file is deleted concurrently by other threads.
+ *
+ * @param jobId ID of the job this blob belongs to
+ * @param blobKey BLOB key associated with the requested file
+ * @return The content of the BLOB.
+ * @throws java.io.FileNotFoundException if the BLOB does not exist;
+ * @throws IOException if any other error occurs when retrieving the file.
+ */
+@Override
+public byte[] readFile(JobID jobId, PermanentBlobKey blobKey) throws 
IOException {
+checkNotNull(jobId);
+checkNotNull(blobKey);
+
+final File localFile = BlobUtils.getStorageLocation(storageDir, jobId, 
blobKey);
+readWriteLock.readLock().lock();
+
+try {
+if (localFile.exists()) {
+blobCacheSizeTracker.update(jobId, blobKey);
+return FileUtils.readAllBytes(localFile.toPath());
+}
+} finally {
+readWriteLock.readLock().unlock();
+}
+
+// first try the distributed blob store (if available)
+// use a temporary file (thread-safe without locking)
+File incomingFile = createTemporaryFilename();
+try {
+try {
+if (blobView.get(jobId, blobKey, incomingFile)) {
+// now move the temp file to our local cache atomically
+readWriteLock.writeLock().lock();
+try {
+checkLimitAndMoveFile(incomingFile, jobId, blobKey, 
localFile, log, null);
+return FileUtils.readAllBytes(localFile.toPath());
+} finally {
+readWriteLock.writeLock().unlock();
+}
+}
+} catch (Exception e) {
+log.info(
+"Failed to copy from blob store. Downloading from BLOB 
server instead.", e);
+}
+
+final InetSocketAddress currentServerAddress = serverAddress;
+
+if (currentServerAddress != null) {
+// fallback: download from the BlobServer
+BlobClient.downloadFromBlobServer(
+jobId,
+blobKey,
+incomingFile,
+currentServerAddress,
+blobClientConfig,
+numFetchRetries);
+
+readWriteLock.writeLock().lock();
+try {
+checkLimitAndMoveFile(incomingFile, jobId, blobKey, 
localFile, log, null);
+return FileUtils.readAllBytes(localFile.toPath());
+} finally {
+readWriteLock.writeLock().unlock();
+}
+} else {
+throw new IOException(
+"Cannot download from BlobServer, because the server 
address is unknown.");
+}
+
+} finally {
+// delete incomingFile from a failed download
+if (!incomingFile.delete() && incomingFile.exists()) {
+log.warn(
+"Could not delete the staging file {} for blob key {} 
and job {}.",
+incomingFile,
+blobKey,
+jobId);
+}
+}
+}
+
+private void checkLimitAndMoveFile(
+File incomingFile,
+@Nullable JobID jobId,
+BlobKey blobKey,
+File localFile,
+Logger log,
+@Nullable BlobStore blobStore)
+throws IOException {
+
+// Check the size limit and delete the files that exceeds the limit
+final long sizeOfIncomingFile = incomingFile.length();
+final List> blobsToDelete =
+blobCacheSizeTracker.checkLimit(sizeOfIncomingFile);
+
+for (Tuple2 key : blobsToDelete) {
+deleteFile(key.f0, key.f1);
+blobCacheSizeTracker.untrack(key);

Review comment:
   >>> The only risk is that we will try 

[GitHub] [flink] zhuzhurk commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


zhuzhurk commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677938458



##
File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/DefaultExecutionGraphDeploymentWithSizeLimitTest.java
##
@@ -0,0 +1,185 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License
+ */
+
+package org.apache.flink.runtime.executiongraph;
+
+import org.apache.flink.configuration.BlobServerOptions;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.runtime.JobException;
+import org.apache.flink.runtime.blob.BlobCacheSizeTracker;
+import org.apache.flink.runtime.blob.BlobServer;
+import org.apache.flink.runtime.blob.PermanentBlobCache;
+import org.apache.flink.runtime.blob.VoidBlobStore;
+import org.apache.flink.runtime.client.JobExecutionException;
+import 
org.apache.flink.runtime.concurrent.ComponentMainThreadExecutorServiceAdapter;
+import org.apache.flink.runtime.deployment.InputGateDeploymentDescriptor;
+import org.apache.flink.runtime.deployment.TaskDeploymentDescriptor;
+import org.apache.flink.runtime.execution.ExecutionState;
+import 
org.apache.flink.runtime.executiongraph.utils.SimpleAckingTaskManagerGateway;
+import org.apache.flink.runtime.io.network.partition.ResultPartitionType;
+import org.apache.flink.runtime.jobgraph.DistributionPattern;
+import org.apache.flink.runtime.jobgraph.IntermediateResultPartitionID;
+import org.apache.flink.runtime.jobgraph.JobGraph;
+import org.apache.flink.runtime.jobgraph.JobGraphTestUtils;
+import org.apache.flink.runtime.jobgraph.JobVertex;
+import org.apache.flink.runtime.jobgraph.JobVertexID;
+import org.apache.flink.runtime.jobmaster.LogicalSlot;
+import org.apache.flink.runtime.jobmaster.TestingLogicalSlotBuilder;
+import org.apache.flink.runtime.operators.BatchTask;
+import org.apache.flink.runtime.scheduler.strategy.ConsumedPartitionGroup;
+import org.apache.flink.runtime.testutils.DirectScheduledExecutorService;
+import org.apache.flink.util.function.FunctionUtils;
+
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
+
+import java.io.IOException;
+import java.net.InetSocketAddress;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.concurrent.ArrayBlockingQueue;
+import java.util.concurrent.BlockingQueue;
+
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNotNull;
+
+public class DefaultExecutionGraphDeploymentWithSizeLimitTest
+extends DefaultExecutionGraphDeploymentWithBlobCacheTest {
+
+@Before
+@Override
+public void setupBlobServer() throws IOException {
+Configuration config = new Configuration();
+// Always offload the serialized JobInformation, TaskInformation and 
cached
+// ShuffleDescriptors
+config.setInteger(BlobServerOptions.OFFLOAD_MINSIZE, 0);
+blobServer = new BlobServer(config, new VoidBlobStore());
+blobServer.start();
+blobWriter = blobServer;
+
+InetSocketAddress serverAddress = new InetSocketAddress("localhost", 
blobServer.getPort());
+// Set the size limit of the blob cache to 1
+BlobCacheSizeTracker blobCacheSizeTracker = new 
BlobCacheSizeTracker(1L);
+blobCache =
+new PermanentBlobCache(
+config, new VoidBlobStore(), serverAddress, 
blobCacheSizeTracker);
+}
+
+@After
+@Override
+public void shutdownBlobServer() throws IOException {
+if (blobServer != null) {
+blobServer.close();
+}
+}
+
+@Test
+public void testDeployTasksWithMinimumSizeLimit() throws Exception {

Review comment:
   Ok. Then how about changing the test name to 
`testDeployMultipleTasksWithBlobCacheSizeLimit`?




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #16613: [BP-1.13][FLINK-23418][e2e] Increase the timeout to make kubernetes application ha test more stable

2021-07-27 Thread GitBox


flinkbot commented on pull request #16613:
URL: https://github.com/apache/flink/pull/16613#issuecomment-887972126


   
   ## CI report:
   
   * 51a7fc3d1abada9a68e4249da55853c5fa1cc1e2 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16612: [FLINK-21207] Fix source table with 'csv.disable-quote-character' = 'true' can not take effect for old csv format version

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16612:
URL: https://github.com/apache/flink/pull/16612#issuecomment-887960095


   
   ## CI report:
   
   * a855139796eca90005c16c59b6a0c8a8159229fe Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21061)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16599: [FLINK-23479][table-planner] Fix the unstable test cases about json plan

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16599:
URL: https://github.com/apache/flink/pull/16599#issuecomment-886750468


   
   ## CI report:
   
   * 8a723c48d4bdd4bcdb543b9ff9e657d4d030280c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21017)
 
   * 976d562df5dd98f70ea407814e487d343ea252b0 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21059)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16611: [FLINK-22911][python] Align FLIP-136 (Improve interoperability between DataStream and Table API) in PyFlink Table API

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16611:
URL: https://github.com/apache/flink/pull/16611#issuecomment-887695133


   
   ## CI report:
   
   * 8d5954945c4eeb370bc8b3f5da3f8356b00f8278 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21047)
 
   * 7ad83d88cfd5c18e857af139b24faa7e7bcf77b1 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21060)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * cbf860c54288c7da244fd439353bf275440b00c7 UNKNOWN
   * 80242347a40dc2e7d1baa2a711d312d541780991 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21056)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-23521) flink sql is not in state running but finished instead aborting checkpoint

2021-07-27 Thread didiaode18 (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

didiaode18 updated FLINK-23521:
---
Summary: flink sql is not in state running but finished instead aborting 
checkpoint  (was: flinksql )

> flink sql is not in state running but finished instead aborting checkpoint
> --
>
> Key: FLINK-23521
> URL: https://issues.apache.org/jira/browse/FLINK-23521
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Scala
>Affects Versions: 1.11.2
>Reporter: didiaode18
>Priority: Major
>
> When I look at Webui, I find that the code table related part of my program 
> does exist with the status of finished.How can I handle it to get checkpoint 
> back to normal? please. thanx
> the same as  
> [https://stackoverflow.com/questions/68004194/flink-sql-is-not-in-state-running-but-finished-instead-aborting-checkpoint]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-23515) I try to specify savepoint dir in my app. However, the startup task reports an error, java.lang.unsupported operationexception: the configuration is unmodifiable; its

2021-07-27 Thread Yun Tang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-23515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17388413#comment-17388413
 ] 

Yun Tang commented on FLINK-23515:
--

You should link the duplicated ticket in this ticket if you close this with 
duplicated information.

As I said, this is a bug caused by FLINK-19466. The {{config}} to set is from 
{{getEnvironment().getTaskManagerInfo().getConfiguration()}} which is a 
{{UnmodifiableConfiguration}}.

> I try to specify savepoint dir in my app. However, the startup task reports 
> an error, java.lang.unsupported operationexception: the configuration is 
> unmodifiable; its contents cannot be changed.
> --
>
> Key: FLINK-23515
> URL: https://issues.apache.org/jira/browse/FLINK-23515
> Project: Flink
>  Issue Type: Bug
>  Components: API / State Processor
>Affects Versions: 1.13.1
> Environment: Flink Version :1.13.1
> Standalone Cluster
>Reporter: 李伟高
>Priority: Major
>
> The directory where savepoint is configured in my application
> {code:java}
> String savepointDir;
> if (Strings.isNotBlank(savepointDir = 
> parameterTool.get("system.savepoint.dir"))) {
> env.setDefaultSavepointDirectory(savepointDir);
> }
> {code}
> Where `system.savepoint.dir` is not empty.
>  
> When I submit an application to my standalone cluster, an error is prompted:
> {code:java}
> 2021-07-27 16:26:34.029 [travel_spilt -> (Sink: sink_es_insert, Sink: 
> sink_es_update) (1/1)#31] WARN  org.apache.flink.runtime.taskmanager.Task  - 
> travel_spilt -> (Sink: sink_es_insert, Sink: sink_es_update) (1/1)#31 
> (1f33925046e04307ba6beb3959d2cad5) switched from DEPLOYING to FAILED with 
> failure cause: java.lang.UnsupportedOperationException: The configuration is 
> unmodifiable; its contents cannot be changed.2021-07-27 16:26:34.029 
> [travel_spilt -> (Sink: sink_es_insert, Sink: sink_es_update) (1/1)#31] WARN  
> org.apache.flink.runtime.taskmanager.Task  - travel_spilt -> (Sink: 
> sink_es_insert, Sink: sink_es_update) (1/1)#31 
> (1f33925046e04307ba6beb3959d2cad5) switched from DEPLOYING to FAILED with 
> failure cause: java.lang.UnsupportedOperationException: The configuration is 
> unmodifiable; its contents cannot be changed. at 
> org.apache.flink.configuration.UnmodifiableConfiguration.error(UnmodifiableConfiguration.java:73)
>  at 
> org.apache.flink.configuration.UnmodifiableConfiguration.setValueInternal(UnmodifiableConfiguration.java:63)
>  at org.apache.flink.configuration.Configuration.set(Configuration.java:709) 
> at 
> org.apache.flink.runtime.state.CheckpointStorageLoader.load(CheckpointStorageLoader.java:174)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.createCheckpointStorage(StreamTask.java:1231)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:337)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:308)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:281)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:273)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:263)
>  at 
> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask.(OneInputStreamTask.java:65)
>  at sun.reflect.GeneratedConstructorAccessor339.newInstance(Unknown Source) 
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>  at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at 
> org.apache.flink.runtime.taskmanager.Task.loadAndInstantiateInvokable(Task.java:1524)
>  at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:730) at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at 
> java.lang.Thread.run(Thread.java:748)
> {code}
> Every stateful operator has this error.What do I need to do? I looked at 
> other versions and didn't find the method Unmodifiable 
> configuration#setvalueinternal (string key, t value).
> Looking forward to your reply, thank you.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-23521) flinksql

2021-07-27 Thread didiaode18 (Jira)
didiaode18 created FLINK-23521:
--

 Summary: flinksql 
 Key: FLINK-23521
 URL: https://issues.apache.org/jira/browse/FLINK-23521
 Project: Flink
  Issue Type: Improvement
  Components: API / Scala
Affects Versions: 1.11.2
Reporter: didiaode18


When I look at Webui, I find that the code table related part of my program 
does exist with the status of finished.How can I handle it to get checkpoint 
back to normal? please. thanx

the same as  
[https://stackoverflow.com/questions/68004194/flink-sql-is-not-in-state-running-but-finished-instead-aborting-checkpoint]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-22860) Supplement 'HELP' command prompt message for SQL-Cli.

2021-07-27 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-22860.
---
Fix Version/s: 1.14
   Resolution: Fixed

Fixed in master: 6aac8560b68f1b7c70e6d6f3356ef877e2eea370

> Supplement 'HELP' command prompt message for SQL-Cli.
> -
>
> Key: FLINK-22860
> URL: https://issues.apache.org/jira/browse/FLINK-22860
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Client
>Reporter: Roc Marshal
>Assignee: Roc Marshal
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.14
>
> Attachments: attach.png
>
>
> !attach.png!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong merged pull request #16060: [FLINK-22860][table sql-client] Supplement 'HELP' command prompt message for SQL-Cli.

2021-07-27 Thread GitBox


wuchong merged pull request #16060:
URL: https://github.com/apache/flink/pull/16060


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] hackergin commented on a change in pull request #16588: [FLINK-22936] Support column comment in Schema and ResolvedSchema

2021-07-27 Thread GitBox


hackergin commented on a change in pull request #16588:
URL: https://github.com/apache/flink/pull/16588#discussion_r677930660



##
File path: 
flink-table/flink-table-common/src/main/java/org/apache/flink/table/catalog/Column.java
##
@@ -108,6 +119,13 @@ public String asSummaryString() {
 sb.append(" ");
 sb.append(e);
 });
+getComment()
+.ifPresent(
+c -> {
+sb.append(" COMMENT '");
+sb.append(c);

Review comment:
   @Airblader  thanks for the review, I update the code, and add the 
related test case. 




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #16613: [BP-1.13][FLINK-23418][e2e] Increase the timeout to make kubernetes application ha test more stable

2021-07-27 Thread GitBox


flinkbot commented on pull request #16613:
URL: https://github.com/apache/flink/pull/16613#issuecomment-887963804


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 51a7fc3d1abada9a68e4249da55853c5fa1cc1e2 (Wed Jul 28 
02:35:53 UTC 2021)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on a change in pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


zhuzhurk commented on a change in pull request #16498:
URL: https://github.com/apache/flink/pull/16498#discussion_r677928575



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/blob/PermanentBlobService.java
##
@@ -50,13 +50,20 @@
 /**
  * Returns the content of the file for the BLOB with the provided job ID 
the blob key.
  *
+ * Compared to {@code getFile}, {@code readFile} will attempt to read 
the entire file after
+ * retrieving it. If file reading and file retrieving is done in the same 
WRITE lock, it can
+ * avoid the scenario that the path to the file is deleted concurrently by 
other threads at the
+ * same time as the file is retrieved and read.

Review comment:
   maybe "at the same time as the file is retrieved and read" -> "when the 
file is retrieved but not read yet"?




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wangyang0918 opened a new pull request #16613: [BP-1.13][FLINK-23418][e2e] Increase the timeout to make kubernetes application ha test more stable

2021-07-27 Thread GitBox


wangyang0918 opened a new pull request #16613:
URL: https://github.com/apache/flink/pull/16613


   Backport #16602 to release-1.13.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-23515) I try to specify savepoint dir in my app. However, the startup task reports an error, java.lang.unsupported operationexception: the configuration is unmodifiable; its

2021-07-27 Thread Jira


[ 
https://issues.apache.org/jira/browse/FLINK-23515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17388400#comment-17388400
 ] 

李伟高 commented on FLINK-23515:
-

[~yunta] Thank you for your reply.

First, repeat the shutdown. Two submissions of issues were caused by my 
network. I closed one of them myself.

Second, I have a question, that is, the source code annotation is that I can 
modify it, and its priority is higher than my flink-conf.yml.Source code in 
rg.apache.flink.runtime.state.CheckpointStorageLoader.class:
{code:java}
if (defaultSavepointDirectory != null) {
// If a savepoint directory was manually specified in code
// we override any value set in the flink-conf. This allows
// us to pass this value to the CheckpointStorage instance
// where it is needed at runtime while keeping its API logically
// separated for users.
config.set(
CheckpointingOptions.SAVEPOINT_DIRECTORY, 
defaultSavepointDirectory.toString());
}
{code}

> I try to specify savepoint dir in my app. However, the startup task reports 
> an error, java.lang.unsupported operationexception: the configuration is 
> unmodifiable; its contents cannot be changed.
> --
>
> Key: FLINK-23515
> URL: https://issues.apache.org/jira/browse/FLINK-23515
> Project: Flink
>  Issue Type: Bug
>  Components: API / State Processor
>Affects Versions: 1.13.1
> Environment: Flink Version :1.13.1
> Standalone Cluster
>Reporter: 李伟高
>Priority: Major
>
> The directory where savepoint is configured in my application
> {code:java}
> String savepointDir;
> if (Strings.isNotBlank(savepointDir = 
> parameterTool.get("system.savepoint.dir"))) {
> env.setDefaultSavepointDirectory(savepointDir);
> }
> {code}
> Where `system.savepoint.dir` is not empty.
>  
> When I submit an application to my standalone cluster, an error is prompted:
> {code:java}
> 2021-07-27 16:26:34.029 [travel_spilt -> (Sink: sink_es_insert, Sink: 
> sink_es_update) (1/1)#31] WARN  org.apache.flink.runtime.taskmanager.Task  - 
> travel_spilt -> (Sink: sink_es_insert, Sink: sink_es_update) (1/1)#31 
> (1f33925046e04307ba6beb3959d2cad5) switched from DEPLOYING to FAILED with 
> failure cause: java.lang.UnsupportedOperationException: The configuration is 
> unmodifiable; its contents cannot be changed.2021-07-27 16:26:34.029 
> [travel_spilt -> (Sink: sink_es_insert, Sink: sink_es_update) (1/1)#31] WARN  
> org.apache.flink.runtime.taskmanager.Task  - travel_spilt -> (Sink: 
> sink_es_insert, Sink: sink_es_update) (1/1)#31 
> (1f33925046e04307ba6beb3959d2cad5) switched from DEPLOYING to FAILED with 
> failure cause: java.lang.UnsupportedOperationException: The configuration is 
> unmodifiable; its contents cannot be changed. at 
> org.apache.flink.configuration.UnmodifiableConfiguration.error(UnmodifiableConfiguration.java:73)
>  at 
> org.apache.flink.configuration.UnmodifiableConfiguration.setValueInternal(UnmodifiableConfiguration.java:63)
>  at org.apache.flink.configuration.Configuration.set(Configuration.java:709) 
> at 
> org.apache.flink.runtime.state.CheckpointStorageLoader.load(CheckpointStorageLoader.java:174)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.createCheckpointStorage(StreamTask.java:1231)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:337)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:308)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:281)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:273)
>  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.(StreamTask.java:263)
>  at 
> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask.(OneInputStreamTask.java:65)
>  at sun.reflect.GeneratedConstructorAccessor339.newInstance(Unknown Source) 
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>  at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at 
> org.apache.flink.runtime.taskmanager.Task.loadAndInstantiateInvokable(Task.java:1524)
>  at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:730) at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) at 
> java.lang.Thread.run(Thread.java:748)
> {code}
> Every stateful operator has this error.What do I need to do? I looked at 
> other versions and didn't find the method Unmodifiable 
> configuration#setvalueinternal (string key, t value).
> Looking forward to your reply, thank you.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-23418) 'Run kubernetes application HA test' fail on Azure

2021-07-27 Thread Yang Wang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23418?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yang Wang updated FLINK-23418:
--
Fix Version/s: 1.13.2

> 'Run kubernetes application HA test' fail on Azure
> --
>
> Key: FLINK-23418
> URL: https://issues.apache.org/jira/browse/FLINK-23418
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes, Runtime / Coordination
>Affects Versions: 1.13.1
>Reporter: Dawid Wysakowicz
>Assignee: Yang Wang
>Priority: Major
>  Labels: pull-request-available, test-stability
> Fix For: 1.14.0, 1.13.2
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=20589=logs=68a897ab-3047-5660-245a-cce8f83859f6=16ca2cca-2f63-5cce-12d2-d519b930a729=3747
> {code}
> Jul 16 23:58:49   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:208)
>  ~[flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:158)
>  ~[flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.actor.Actor$class.aroundReceive(Actor.scala:517) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at akka.actor.ActorCell.invoke(ActorCell.scala:561) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at akka.dispatch.Mailbox.run(Mailbox.scala:225) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at akka.dispatch.Mailbox.exec(Mailbox.scala:235) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) 
> [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>  [flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49 Caused by: akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka.tcp://flink@172.17.0.3:6123/user/rpc/jobmanager_2#2101744934]] 
> after [1 ms]. Message of type 
> [org.apache.flink.runtime.rpc.messages.RemoteFencedMessage]. A typical reason 
> for `AskTimeoutException` is that the recipient actor didn't send a reply.
> Jul 16 23:58:49   at 
> akka.pattern.PromiseActorRef$$anonfun$2.apply(AskSupport.scala:635) 
> ~[flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.pattern.PromiseActorRef$$anonfun$2.apply(AskSupport.scala:635) 
> ~[flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:648) 
> ~[flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> akka.actor.Scheduler$$anon$4.run(Scheduler.scala:205) 
> ~[flink-dist_2.11-1.13-SNAPSHOT.jar:1.13-SNAPSHOT]
> Jul 16 23:58:49   at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
>  

[GitHub] [flink] flinkbot edited a comment on pull request #16611: [FLINK-22911][python] Align FLIP-136 (Improve interoperability between DataStream and Table API) in PyFlink Table API

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16611:
URL: https://github.com/apache/flink/pull/16611#issuecomment-887695133


   
   ## CI report:
   
   * 8d5954945c4eeb370bc8b3f5da3f8356b00f8278 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21047)
 
   * 7ad83d88cfd5c18e857af139b24faa7e7bcf77b1 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #16612: [FLINK-21207] Fix source table with 'csv.disable-quote-character' = 'true' can not take effect for old csv format version

2021-07-27 Thread GitBox


flinkbot commented on pull request #16612:
URL: https://github.com/apache/flink/pull/16612#issuecomment-887960095


   
   ## CI report:
   
   * a855139796eca90005c16c59b6a0c8a8159229fe UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16599: [FLINK-23479][table-planner] Fix the unstable test cases about json plan

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16599:
URL: https://github.com/apache/flink/pull/16599#issuecomment-886750468


   
   ## CI report:
   
   * 8a723c48d4bdd4bcdb543b9ff9e657d4d030280c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21017)
 
   * 976d562df5dd98f70ea407814e487d343ea252b0 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Comment Edited] (FLINK-23263) LocalBufferPool can not request memory.

2021-07-27 Thread Ada Wong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-23263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17388397#comment-17388397
 ] 

Ada Wong edited comment on FLINK-23263 at 7/28/21, 2:12 AM:


This happen in double stream join and I don't set ttl. OutputBuffer of join 
operator is filled. Flink can't consume any kafka records, when outputBuffer of 
join operator is filled. 


was (Author: ana4):
This happen in double stream join and I don't set ttl. OutputBuffer of join 
operator is filled. Flink can't consume any kafka records, when outputBuffer of 
join operator is filled. 

> LocalBufferPool can not request memory.
> ---
>
> Key: FLINK-23263
> URL: https://issues.apache.org/jira/browse/FLINK-23263
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.1
>Reporter: Ada Wong
>Priority: Major
>
> Flink job is running, bug it can not consume kafka data.
> This following is exception.
> "Source: test_records (2/3)" #78 prio=5 os_prio=0 tid=0x7fd4c4a24800 
> nid=0x21bf in Object.wait() [0x7fd4d581a000]
>java.lang.Thread.State: TIMED_WAITING (on object monitor)
>   at java.lang.Object.wait(Native Method)
>   at 
> org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestMemorySegment(LocalBufferPool.java:251)
>   - locked <0x00074d8b0df0> (a java.util.ArrayDeque)
>   at 
> org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBufferBuilderBlocking(LocalBufferPool.java:218)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.requestNewBufferBuilder(RecordWriter.java:264)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.copyFromSerializerToTargetChannel(RecordWriter.java:192)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:162)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:128)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:107)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:89)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:45)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:718)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:696)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collect(StreamSourceContexts.java:104)
>   - locked <0x00074cbd3be0> (a java.lang.Object)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collectWithTimestamp(StreamSourceContexts.java:111)
>   at 
> org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordWithTimestamp(AbstractFetcher.java:398)
>   - locked <0x00074cbd3be0> (a java.lang.Object)
>   at 
> org.apache.flink.streaming.connectors.kafka.internal.Kafka010Fetcher.emitRecord(Kafka010Fetcher.java:91)
>   at 
> org.apache.flink.streaming.connectors.kafka.internal.Kafka09Fetcher.runFetchLoop(Kafka09Fetcher.java:156)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:711)
>   at 
> com.dtstack.flink.sql.source.kafka.KafkaConsumer011.run(KafkaConsumer011.java:66)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:93)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:57)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:97)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:302)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>   at java.lang.Thread.run(Thread.java:748)
> "Map -> to: Tuple2 -> Map -> (from: (id, sid, item, val, unit, dt,
> after_index, tablename, PROCTIME) -> where: (AND(=(tablename,
> CONCAT(_UTF-16LE't_real', currtime2(dt, _UTF-16LE'MMdd'))),
> OR(=(after_index, _UTF-16LE'2.6'), AND(=(sid, _UTF-16LE'7296'),
> =(after_index, _UTF-16LE'2.10'))), LIKE(item, _UTF-16LE'??5%'),
> =(MOD(getminutes(dt), 5), 0))), select: (id AS ID, sid AS STID, item
> AS ITEM, val AS VAL, unit AS UNIT, dt AS DATATIME), from: (id, sid,
> item, val, unit, dt, after_index, tablename, PROCTIME) -> where:
> (AND(=(tablename, CONCAT(_UTF-16LE't_real', currtime2(dt,
> _UTF-16LE'MMdd'))), OR(=(after_index, _UTF-16LE'2.6'), AND(=(sid,
> _UTF-16LE'7296'), =(after_index, _UTF-16LE'2.10'))), LIKE(item,
> 

[jira] [Commented] (FLINK-23263) LocalBufferPool can not request memory.

2021-07-27 Thread Ada Wong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-23263?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17388397#comment-17388397
 ] 

Ada Wong commented on FLINK-23263:
--

This happen in double stream join and I don't set ttl. OutputBuffer of join 
operator is filled. Flink can't consume any kafka records, when outputBuffer of 
join operator is filled. 

> LocalBufferPool can not request memory.
> ---
>
> Key: FLINK-23263
> URL: https://issues.apache.org/jira/browse/FLINK-23263
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.1
>Reporter: Ada Wong
>Priority: Major
>
> Flink job is running, bug it can not consume kafka data.
> This following is exception.
> "Source: test_records (2/3)" #78 prio=5 os_prio=0 tid=0x7fd4c4a24800 
> nid=0x21bf in Object.wait() [0x7fd4d581a000]
>java.lang.Thread.State: TIMED_WAITING (on object monitor)
>   at java.lang.Object.wait(Native Method)
>   at 
> org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestMemorySegment(LocalBufferPool.java:251)
>   - locked <0x00074d8b0df0> (a java.util.ArrayDeque)
>   at 
> org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBufferBuilderBlocking(LocalBufferPool.java:218)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.requestNewBufferBuilder(RecordWriter.java:264)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.copyFromSerializerToTargetChannel(RecordWriter.java:192)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:162)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:128)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:107)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:89)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:45)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:718)
>   at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator$CountingOutput.collect(AbstractStreamOperator.java:696)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collect(StreamSourceContexts.java:104)
>   - locked <0x00074cbd3be0> (a java.lang.Object)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$NonTimestampContext.collectWithTimestamp(StreamSourceContexts.java:111)
>   at 
> org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordWithTimestamp(AbstractFetcher.java:398)
>   - locked <0x00074cbd3be0> (a java.lang.Object)
>   at 
> org.apache.flink.streaming.connectors.kafka.internal.Kafka010Fetcher.emitRecord(Kafka010Fetcher.java:91)
>   at 
> org.apache.flink.streaming.connectors.kafka.internal.Kafka09Fetcher.runFetchLoop(Kafka09Fetcher.java:156)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:711)
>   at 
> com.dtstack.flink.sql.source.kafka.KafkaConsumer011.run(KafkaConsumer011.java:66)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:93)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:57)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:97)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:302)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711)
>   at java.lang.Thread.run(Thread.java:748)
> "Map -> to: Tuple2 -> Map -> (from: (id, sid, item, val, unit, dt,
> after_index, tablename, PROCTIME) -> where: (AND(=(tablename,
> CONCAT(_UTF-16LE't_real', currtime2(dt, _UTF-16LE'MMdd'))),
> OR(=(after_index, _UTF-16LE'2.6'), AND(=(sid, _UTF-16LE'7296'),
> =(after_index, _UTF-16LE'2.10'))), LIKE(item, _UTF-16LE'??5%'),
> =(MOD(getminutes(dt), 5), 0))), select: (id AS ID, sid AS STID, item
> AS ITEM, val AS VAL, unit AS UNIT, dt AS DATATIME), from: (id, sid,
> item, val, unit, dt, after_index, tablename, PROCTIME) -> where:
> (AND(=(tablename, CONCAT(_UTF-16LE't_real', currtime2(dt,
> _UTF-16LE'MMdd'))), OR(=(after_index, _UTF-16LE'2.6'), AND(=(sid,
> _UTF-16LE'7296'), =(after_index, _UTF-16LE'2.10'))), LIKE(item,
> _UTF-16LE'??5%'), =(MOD(getminutes(dt), 5), 0))), select: (sid AS
> STID, val AS VAL, dt AS DATATIME)) (1/1)" #79 prio=5 os_prio=0
> tid=0x7fd4c4a94000 nid=0x21c0 in Object.wait()
> [0x7fd4d5719000]
> java.lang.Thread.State: TIMED_WAITING (on object monitor)
> 

[jira] [Commented] (FLINK-19739) CompileException when windowing in batch mode: A method named "replace" is not declared in any enclosing class nor any supertype

2021-07-27 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17388395#comment-17388395
 ] 

Jingsong Lee commented on FLINK-19739:
--

[~TsReaper] Can you cherry-pick to 1.13?

> CompileException when windowing in batch mode: A method named "replace" is 
> not declared in any enclosing class nor any supertype 
> -
>
> Key: FLINK-19739
> URL: https://issues.apache.org/jira/browse/FLINK-19739
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.2, 1.12.0
> Environment: Ubuntu 18.04
> Python 3.8, jar built from master yesterday.
> Or Python 3.7, installed latest version from pip.
>Reporter: Alex Hall
>Assignee: Caizhi Weng
>Priority: Minor
>  Labels: auto-deprioritized-major, auto-unassigned, 
> pull-request-available
>
> Example script:
> {code:python}
> from pyflink.table import EnvironmentSettings, BatchTableEnvironment
> from pyflink.table.window import Tumble
> env_settings = (
> 
> EnvironmentSettings.new_instance().in_batch_mode().use_blink_planner().build()
> )
> table_env = BatchTableEnvironment.create(environment_settings=env_settings)
> table_env.execute_sql(
> """
> CREATE TABLE table1 (
> amount INT,
> ts TIMESTAMP(3),
> WATERMARK FOR ts AS ts - INTERVAL '5' SECOND
> ) WITH (
> 'connector.type' = 'filesystem',
> 'format.type' = 'csv',
> 'connector.path' = '/home/alex/work/test-flink/data1.csv'
> )
> """
> )
> table1 = table_env.from_path("table1")
> table = (
> table1
> .window(Tumble.over("5.days").on("ts").alias("__window"))
> .group_by("__window")
> .select("amount.sum")
> )
> print(table.to_pandas())
> {code}
> Output:
> {code}
> WARNING: An illegal reflective access operation has occurred
> WARNING: Illegal reflective access by 
> org.apache.flink.api.python.shaded.io.netty.util.internal.ReflectionUtil 
> (file:/home/alex/work/flink/flink-dist/target/flink-1.12-SNAPSHOT-bin/flink-1.12-SNAPSHOT/opt/flink-python_2.11-1.12-SNAPSHOT.jar)
>  to constructor java.nio.DirectByteBuffer(long,int)
> WARNING: Please consider reporting this to the maintainers of 
> org.apache.flink.api.python.shaded.io.netty.util.internal.ReflectionUtil
> WARNING: Use --illegal-access=warn to enable warnings of further illegal 
> reflective access operations
> WARNING: All illegal access operations will be denied in a future release
> /* 1 */
> /* 2 */  public class LocalHashWinAggWithoutKeys$59 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */  implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator, 
> org.apache.flink.streaming.api.operators.BoundedOneInput {
> /* 4 */
> /* 5 */private final Object[] references;
> /* 6 */
> /* 7 */private static final org.slf4j.Logger LOG$2 =
> /* 8 */  org.slf4j.LoggerFactory.getLogger("LocalHashWinAgg");
> /* 9 */
> /* 10 */private transient 
> org.apache.flink.table.types.logical.LogicalType[] aggMapKeyTypes$5;
> /* 11 */private transient 
> org.apache.flink.table.types.logical.LogicalType[] aggBufferTypes$6;
> /* 12 */private transient 
> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap 
> aggregateMap$7;
> /* 13 */org.apache.flink.table.data.binary.BinaryRowData 
> emptyAggBuffer$9 = new org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 14 */org.apache.flink.table.data.writer.BinaryRowWriter 
> emptyAggBufferWriterTerm$10 = new 
> org.apache.flink.table.data.writer.BinaryRowWriter(emptyAggBuffer$9);
> /* 15 */org.apache.flink.table.data.GenericRowData hashAggOutput = 
> new org.apache.flink.table.data.GenericRowData(2);
> /* 16 */private transient 
> org.apache.flink.table.data.binary.BinaryRowData reuseAggMapKey$17 = new 
> org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 17 */private transient 
> org.apache.flink.table.data.binary.BinaryRowData reuseAggBuffer$18 = new 
> org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 18 */private transient 
> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap.Entry 
> reuseAggMapEntry$19 = new 
> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap.Entry(reuseAggMapKey$17,
>  reuseAggBuffer$18);
> /* 19 */org.apache.flink.table.data.binary.BinaryRowData aggMapKey$3 
> = new org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 20 */org.apache.flink.table.data.writer.BinaryRowWriter 
> aggMapKeyWriter$4 = new 
> org.apache.flink.table.data.writer.BinaryRowWriter(aggMapKey$3);
> /* 21 */private 

[jira] [Assigned] (FLINK-19739) CompileException when windowing in batch mode: A method named "replace" is not declared in any enclosing class nor any supertype

2021-07-27 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee reassigned FLINK-19739:


Assignee: Caizhi Weng

> CompileException when windowing in batch mode: A method named "replace" is 
> not declared in any enclosing class nor any supertype 
> -
>
> Key: FLINK-19739
> URL: https://issues.apache.org/jira/browse/FLINK-19739
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.2, 1.12.0
> Environment: Ubuntu 18.04
> Python 3.8, jar built from master yesterday.
> Or Python 3.7, installed latest version from pip.
>Reporter: Alex Hall
>Assignee: Caizhi Weng
>Priority: Minor
>  Labels: auto-deprioritized-major, auto-unassigned, 
> pull-request-available
>
> Example script:
> {code:python}
> from pyflink.table import EnvironmentSettings, BatchTableEnvironment
> from pyflink.table.window import Tumble
> env_settings = (
> 
> EnvironmentSettings.new_instance().in_batch_mode().use_blink_planner().build()
> )
> table_env = BatchTableEnvironment.create(environment_settings=env_settings)
> table_env.execute_sql(
> """
> CREATE TABLE table1 (
> amount INT,
> ts TIMESTAMP(3),
> WATERMARK FOR ts AS ts - INTERVAL '5' SECOND
> ) WITH (
> 'connector.type' = 'filesystem',
> 'format.type' = 'csv',
> 'connector.path' = '/home/alex/work/test-flink/data1.csv'
> )
> """
> )
> table1 = table_env.from_path("table1")
> table = (
> table1
> .window(Tumble.over("5.days").on("ts").alias("__window"))
> .group_by("__window")
> .select("amount.sum")
> )
> print(table.to_pandas())
> {code}
> Output:
> {code}
> WARNING: An illegal reflective access operation has occurred
> WARNING: Illegal reflective access by 
> org.apache.flink.api.python.shaded.io.netty.util.internal.ReflectionUtil 
> (file:/home/alex/work/flink/flink-dist/target/flink-1.12-SNAPSHOT-bin/flink-1.12-SNAPSHOT/opt/flink-python_2.11-1.12-SNAPSHOT.jar)
>  to constructor java.nio.DirectByteBuffer(long,int)
> WARNING: Please consider reporting this to the maintainers of 
> org.apache.flink.api.python.shaded.io.netty.util.internal.ReflectionUtil
> WARNING: Use --illegal-access=warn to enable warnings of further illegal 
> reflective access operations
> WARNING: All illegal access operations will be denied in a future release
> /* 1 */
> /* 2 */  public class LocalHashWinAggWithoutKeys$59 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */  implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator, 
> org.apache.flink.streaming.api.operators.BoundedOneInput {
> /* 4 */
> /* 5 */private final Object[] references;
> /* 6 */
> /* 7 */private static final org.slf4j.Logger LOG$2 =
> /* 8 */  org.slf4j.LoggerFactory.getLogger("LocalHashWinAgg");
> /* 9 */
> /* 10 */private transient 
> org.apache.flink.table.types.logical.LogicalType[] aggMapKeyTypes$5;
> /* 11 */private transient 
> org.apache.flink.table.types.logical.LogicalType[] aggBufferTypes$6;
> /* 12 */private transient 
> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap 
> aggregateMap$7;
> /* 13 */org.apache.flink.table.data.binary.BinaryRowData 
> emptyAggBuffer$9 = new org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 14 */org.apache.flink.table.data.writer.BinaryRowWriter 
> emptyAggBufferWriterTerm$10 = new 
> org.apache.flink.table.data.writer.BinaryRowWriter(emptyAggBuffer$9);
> /* 15 */org.apache.flink.table.data.GenericRowData hashAggOutput = 
> new org.apache.flink.table.data.GenericRowData(2);
> /* 16 */private transient 
> org.apache.flink.table.data.binary.BinaryRowData reuseAggMapKey$17 = new 
> org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 17 */private transient 
> org.apache.flink.table.data.binary.BinaryRowData reuseAggBuffer$18 = new 
> org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 18 */private transient 
> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap.Entry 
> reuseAggMapEntry$19 = new 
> org.apache.flink.table.runtime.operators.aggregate.BytesHashMap.Entry(reuseAggMapKey$17,
>  reuseAggBuffer$18);
> /* 19 */org.apache.flink.table.data.binary.BinaryRowData aggMapKey$3 
> = new org.apache.flink.table.data.binary.BinaryRowData(1);
> /* 20 */org.apache.flink.table.data.writer.BinaryRowWriter 
> aggMapKeyWriter$4 = new 
> org.apache.flink.table.data.writer.BinaryRowWriter(aggMapKey$3);
> /* 21 */private boolean hasInput = false;
> /* 22 */

[GitHub] [flink] JingsongLi merged pull request #16591: [FLINK-19739][table-runtime] Fix compile exception for window aggregate in batch jobs

2021-07-27 Thread GitBox


JingsongLi merged pull request #16591:
URL: https://github.com/apache/flink/pull/16591


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Assigned] (FLINK-22911) Align FLIP-136 (Improve interoperability between DataStream and Table API) in PyFlink Table API

2021-07-27 Thread Huang Xingbo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22911?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Huang Xingbo reassigned FLINK-22911:


Assignee: Huang Xingbo

> Align FLIP-136 (Improve interoperability between DataStream and Table API) in 
> PyFlink Table API
> ---
>
> Key: FLINK-22911
> URL: https://issues.apache.org/jira/browse/FLINK-22911
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Python
>Reporter: Dian Fu
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: auto-unassigned, pull-request-available
> Fix For: 1.14.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] yangjunhan commented on pull request #16603: [FLINK-23111][runtime-web] Bump angular's and ng-zorro's version to 8

2021-07-27 Thread GitBox


yangjunhan commented on pull request #16603:
URL: https://github.com/apache/flink/pull/16603#issuecomment-887945954


   > BTW, I'm also totally fine with doing the upgrades all the way to Angular 
12 on the same PR. Just separating by commit to make it easier to review would 
be appreciated, but we don't need to wait on merges every single time that way.
   
   Sure, I will continue to push commits to this PR.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #16612: [FLINK-21207] Fix source table with 'csv.disable-quote-character' = 'true' can not take effect for old csv format version

2021-07-27 Thread GitBox


flinkbot commented on pull request #16612:
URL: https://github.com/apache/flink/pull/16612#issuecomment-887944673


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit a855139796eca90005c16c59b6a0c8a8159229fe (Wed Jul 28 
01:40:40 UTC 2021)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] hehuiyuan opened a new pull request #16612: [FLINK-21207] Fix source table with 'csv.disable-quote-character' = 'true' can not take effect

2021-07-27 Thread GitBox


hehuiyuan opened a new pull request #16612:
URL: https://github.com/apache/flink/pull/16612


   
   ## What is the purpose of the change
   
   Fix source table with 'csv.disable-quote-character' = 'true' can not take 
effect for old csv format version
   
   ## Brief change log
   
   Invoke disableQuoteChar of CsvSchema when 'disable-quote-character' is 
enabled
   
   
   ## Does this pull request potentially affect one of the following parts:
   
   Dependencies (does it add or upgrade a dependency): (no)
   - The public API, i.e., is any changed class annotated with 
@Public(Evolving): (no)
   - The serializers: (no)
   - The runtime per-record code paths (performance sensitive): (no)
   - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, 
   - Zookeeper: (no)
   - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] mans2singh commented on a change in pull request #16587: [FLINK-23490][examples][table] - Updated print output

2021-07-27 Thread GitBox


mans2singh commented on a change in pull request #16587:
URL: https://github.com/apache/flink/pull/16587#discussion_r677896392



##
File path: 
flink-examples/flink-examples-table/src/main/java/org/apache/flink/table/examples/java/basics/StreamWindowSQLExample.java
##
@@ -79,16 +77,16 @@ public static void main(String[] args) throws Exception {
 + "  COUNT(DISTINCT product) unique_products\n"
 + "FROM orders\n"
 + "GROUP BY TUMBLE(ts, INTERVAL '5' SECOND)";
-Table result = tEnv.sqlQuery(query);
-tEnv.toAppendStream(result, Row.class).print();
 
-// after the table program is converted to DataStream program,
-// we must use `env.execute()` to submit the job.
-env.execute("Streaming Window SQL Job");
+tEnv.executeSql(query).print();
+// should output;

Review comment:
   @JingsongLi - I've corrected the code based on your advice.  Please let 
me know you have any additional comments.  Thanks again for your 
review/comments.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] YikSanChan edited a comment on pull request #15813: [FLINK-22519][flink-python] support tar python archives

2021-07-27 Thread GitBox


YikSanChan edited a comment on pull request #15813:
URL: https://github.com/apache/flink/pull/15813#issuecomment-887928353


   > @YikSanChan Any update ?
   
   Hi Jeff, unfortunately I am not able to work on this recently.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] YikSanChan commented on pull request #15813: [FLINK-22519][flink-python] support tar python archives

2021-07-27 Thread GitBox


YikSanChan commented on pull request #15813:
URL: https://github.com/apache/flink/pull/15813#issuecomment-887928353


   > @YikSanChan Any update ?
   
   Hi Jeff, I am not able to work on this recently.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16587: [FLINK-23490][examples][table] - Updated print output

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16587:
URL: https://github.com/apache/flink/pull/16587#issuecomment-886122043


   
   ## CI report:
   
   * 3eedc57632b94d0832cf308b58fa4c2008462f6b UNKNOWN
   * 45e360e6f9fa1c6ea5dfc2a59651c34a54deba4a Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21006)
 
   * f38ef3af364109bffbdbd5cc8b812c40b2af3eb9 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21057)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18634) FlinkKafkaProducerITCase.testRecoverCommittedTransaction failed with "Timeout expired after 60000milliseconds while awaiting InitProducerId"

2021-07-27 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-18634:
---
  Labels: auto-deprioritized-blocker auto-unassigned test-stability  (was: 
auto-unassigned stale-blocker test-stability)
Priority: Critical  (was: Blocker)

This issue was labeled "stale-blocker" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually a Blocker, 
please raise the priority and ask a committer to assign you the issue or revive 
the public discussion.


> FlinkKafkaProducerITCase.testRecoverCommittedTransaction failed with "Timeout 
> expired after 6milliseconds while awaiting InitProducerId"
> 
>
> Key: FLINK-18634
> URL: https://issues.apache.org/jira/browse/FLINK-18634
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Tests
>Affects Versions: 1.11.0, 1.12.0, 1.13.0
>Reporter: Dian Fu
>Priority: Critical
>  Labels: auto-deprioritized-blocker, auto-unassigned, 
> test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4590=logs=c5f0071e-1851-543e-9a45-9ac140befc32=684b1416-4c17-504e-d5ab-97ee44e08a20
> {code}
> 2020-07-17T11:43:47.9693015Z [ERROR] Tests run: 12, Failures: 0, Errors: 1, 
> Skipped: 0, Time elapsed: 269.399 s <<< FAILURE! - in 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase
> 2020-07-17T11:43:47.9693862Z [ERROR] 
> testRecoverCommittedTransaction(org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase)
>   Time elapsed: 60.679 s  <<< ERROR!
> 2020-07-17T11:43:47.9694737Z org.apache.kafka.common.errors.TimeoutException: 
> org.apache.kafka.common.errors.TimeoutException: Timeout expired after 
> 6milliseconds while awaiting InitProducerId
> 2020-07-17T11:43:47.9695376Z Caused by: 
> org.apache.kafka.common.errors.TimeoutException: Timeout expired after 
> 6milliseconds while awaiting InitProducerId
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #16594: [FLINK-23063][python][table] Remove #connect

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16594:
URL: https://github.com/apache/flink/pull/16594#issuecomment-886556323


   
   ## CI report:
   
   * f6870e09386e9c63459c2de0de3220ad6d028e0d UNKNOWN
   * 4b2cae270f8e378cf7ac80a2b262db2b83752023 UNKNOWN
   * ad84d21a38c492d65ba708c222edc295b431da29 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21051)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16587: [FLINK-23490][examples][table] - Updated print output

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16587:
URL: https://github.com/apache/flink/pull/16587#issuecomment-886122043


   
   ## CI report:
   
   * 3eedc57632b94d0832cf308b58fa4c2008462f6b UNKNOWN
   * 45e360e6f9fa1c6ea5dfc2a59651c34a54deba4a Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21006)
 
   * f38ef3af364109bffbdbd5cc8b812c40b2af3eb9 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16600: [FLINK-19929] Upgrade Kinesis dependencies

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16600:
URL: https://github.com/apache/flink/pull/16600#issuecomment-886780305


   
   ## CI report:
   
   * dccbfc290b390865ce848211f7615ef5aed35efe Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21048)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16556: [FLINK-23452][runtime] Added the throughput meter for measurement of …

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16556:
URL: https://github.com/apache/flink/pull/16556#issuecomment-884238774


   
   ## CI report:
   
   * 4033dd607f8ea9d375c48e3054f070221f897674 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21046)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * f2628656933e4204ee708066a9197101cfdbc72b Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21044)
 
   * cbf860c54288c7da244fd439353bf275440b00c7 UNKNOWN
   * 80242347a40dc2e7d1baa2a711d312d541780991 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21056)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16610: [FLINK-23497][table-planner] Add dependency for scala-parser-combinators

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16610:
URL: https://github.com/apache/flink/pull/16610#issuecomment-887678681


   
   ## CI report:
   
   * 80d379567768967b893645deffb6d996bfef9982 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21045)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * 4d6e1fcdba2be3c29fd6bc6c1a0a4c576f91b0b6 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21035)
 
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * f2628656933e4204ee708066a9197101cfdbc72b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21044)
 
   * cbf860c54288c7da244fd439353bf275440b00c7 UNKNOWN
   * 80242347a40dc2e7d1baa2a711d312d541780991 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21056)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Airblader commented on a change in pull request #16588: [FLINK-22936] Support column comment in Schema and ResolvedSchema

2021-07-27 Thread GitBox


Airblader commented on a change in pull request #16588:
URL: https://github.com/apache/flink/pull/16588#discussion_r677766467



##
File path: 
flink-table/flink-table-common/src/main/java/org/apache/flink/table/catalog/Column.java
##
@@ -108,6 +119,13 @@ public String asSummaryString() {
 sb.append(" ");
 sb.append(e);
 });
+getComment()
+.ifPresent(
+c -> {
+sb.append(" COMMENT '");
+sb.append(c);

Review comment:
   We should wrap this with `EncodingUtils#escapeSingleQuotes`.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Airblader commented on pull request #16603: [FLINK-23111][runtime-web] Bump angular's and ng-zorro's version to 8

2021-07-27 Thread GitBox


Airblader commented on pull request #16603:
URL: https://github.com/apache/flink/pull/16603#issuecomment-887801129


   BTW, I'm also totally fine with doing the upgrades all the way to Angular 12 
on the same PR. Just separating by commit to make it easier to review would be 
appreciated, but we don't need to wait on merges every single time that way.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] sjwiesman commented on a change in pull request #244: [FLINK-23520][datastream] Improve StateFun <-> DataStream Interop

2021-07-27 Thread GitBox


sjwiesman commented on a change in pull request #244:
URL: https://github.com/apache/flink-statefun/pull/244#discussion_r677762089



##
File path: 
statefun-flink/statefun-flink-datastream/src/main/java/org/apache/flink/statefun/sdk/TypedValue.java
##
@@ -0,0 +1,70 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.statefun.sdk;
+
+import java.util.Arrays;
+import java.util.Objects;
+import org.apache.flink.api.common.typeinfo.TypeInfo;
+import org.apache.flink.statefun.sdk.types.TypedValueTypeInfoFactory;
+
+/** A typed value containing a {@link TypeName} and serialized bytes. */
+@TypeInfo(TypedValueTypeInfoFactory.class)
+@SuppressWarnings("unused")
+public class TypedValue {

Review comment:
   The purpose of this class is to avoid exposing the generated protobuf 
TypedValue in the public API. Let me know if you believe I'm over thinking 
this. 




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] sjwiesman commented on a change in pull request #244: [FLINK-23520][datastream] Improve StateFun <-> DataStream Interop

2021-07-27 Thread GitBox


sjwiesman commented on a change in pull request #244:
URL: https://github.com/apache/flink-statefun/pull/244#discussion_r677655807



##
File path: 
statefun-flink/statefun-flink-datastream/src/main/java/org/apache/flink/statefun/sdk/RawValue.java
##
@@ -0,0 +1,68 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.statefun.sdk;
+
+import java.util.Arrays;
+import java.util.Objects;
+import org.apache.flink.api.common.typeinfo.TypeInfo;
+import org.apache.flink.statefun.sdk.types.RawValueTypeInfoFactory;
+
+/** A typed value containing a {@link TypeName} and serialized bytes. */
+@TypeInfo(RawValueTypeInfoFactory.class)
+public class RawValue {

Review comment:
   This just to prevent exposing `TypedValue` in the public API. 




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * 4d6e1fcdba2be3c29fd6bc6c1a0a4c576f91b0b6 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21035)
 
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * f2628656933e4204ee708066a9197101cfdbc72b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21044)
 
   * cbf860c54288c7da244fd439353bf275440b00c7 UNKNOWN
   * 80242347a40dc2e7d1baa2a711d312d541780991 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16594: [FLINK-23063][python][table] Remove #connect

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16594:
URL: https://github.com/apache/flink/pull/16594#issuecomment-886556323


   
   ## CI report:
   
   * f6870e09386e9c63459c2de0de3220ad6d028e0d UNKNOWN
   * 4b2cae270f8e378cf7ac80a2b262db2b83752023 UNKNOWN
   * a47808d9b556988301e767eea2731989ec3c4663 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21042)
 
   * ad84d21a38c492d65ba708c222edc295b431da29 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21051)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16611: [FLINK-22911][python] Align FLIP-136 (Improve interoperability between DataStream and Table API) in PyFlink Table API

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16611:
URL: https://github.com/apache/flink/pull/16611#issuecomment-887695133


   
   ## CI report:
   
   * 8d5954945c4eeb370bc8b3f5da3f8356b00f8278 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21047)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16598: [FLINK-23501][GCP PubSub] add table source API for GCP PubSub

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16598:
URL: https://github.com/apache/flink/pull/16598#issuecomment-886655820


   
   ## CI report:
   
   * 65f256e785b3ac499d38eee63d177f6a6059b609 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21052)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * 4d6e1fcdba2be3c29fd6bc6c1a0a4c576f91b0b6 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21035)
 
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * f2628656933e4204ee708066a9197101cfdbc72b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21044)
 
   * cbf860c54288c7da244fd439353bf275440b00c7 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16594: [FLINK-23063][python][table] Remove #connect

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16594:
URL: https://github.com/apache/flink/pull/16594#issuecomment-886556323


   
   ## CI report:
   
   * f6870e09386e9c63459c2de0de3220ad6d028e0d UNKNOWN
   * 4b2cae270f8e378cf7ac80a2b262db2b83752023 UNKNOWN
   * 17cef7e332076311fb1af59272f22d98c8610f11 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21036)
 
   * a47808d9b556988301e767eea2731989ec3c4663 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21042)
 
   * ad84d21a38c492d65ba708c222edc295b431da29 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21051)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16498: [FLINK-23354][blob] Limit the size of ShuffleDescriptors in PermanentBlobCache on TaskExecutor

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16498:
URL: https://github.com/apache/flink/pull/16498#issuecomment-880512519


   
   ## CI report:
   
   * ea075fb62bc7853bcdbbf6d275c70048a9915f57 UNKNOWN
   * 7ce8a08f3d996eee8c4ac7ac2de4618152099ac7 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21038)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16606: [draft] materialization for FLIP 158

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16606:
URL: https://github.com/apache/flink/pull/16606#issuecomment-887431748


   
   ## CI report:
   
   * 1206a78c6480a509dceafaa81d1baaf1272fdeec Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21040)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Assigned] (FLINK-23518) ZooKeeperLeaderElectionConnectionHandlingTest hangs on Azure

2021-07-27 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler reassigned FLINK-23518:


Assignee: Chesnay Schepler

> ZooKeeperLeaderElectionConnectionHandlingTest hangs on Azure
> 
>
> Key: FLINK-23518
> URL: https://issues.apache.org/jira/browse/FLINK-23518
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.14.0
>Reporter: Dawid Wysakowicz
>Assignee: Chesnay Schepler
>Priority: Critical
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=20964=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=24c3384f-1bcb-57b3-224f-51bf973bbee8=9511
> {code}
> Jul 26 11:05:26 "main" #1 prio=5 os_prio=0 tid=0x7efd7000b800 nid=0x1769 
> waiting on condition [0x7efd771ea000]
> Jul 26 11:05:26java.lang.Thread.State: WAITING (parking)
> Jul 26 11:05:26   at sun.misc.Unsafe.park(Native Method)
> Jul 26 11:05:26   - parking to wait for  <0x86681f00> (a 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
> Jul 26 11:05:26   at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> Jul 26 11:05:26   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> Jul 26 11:05:26   at 
> java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:403)
> Jul 26 11:05:26   at 
> org.apache.flink.runtime.leaderelection.ZooKeeperLeaderElectionConnectionHandlingTest$QueueLeaderElectionListener.next(ZooKeeperLeaderElectionConnectionHandlingTest.java:324)
> Jul 26 11:05:26   at 
> org.apache.flink.runtime.leaderelection.ZooKeeperLeaderElectionConnectionHandlingTest$QueueLeaderElectionListener.next(ZooKeeperLeaderElectionConnectionHandlingTest.java:318)
> Jul 26 11:05:26   at 
> org.apache.flink.runtime.leaderelection.ZooKeeperLeaderElectionConnectionHandlingTest.testConnectionSuspendedHandlingDuringInitialization(ZooKeeperLeaderElectionConnectionHandlingTest.java:114)
> Jul 26 11:05:26   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Jul 26 11:05:26   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Jul 26 11:05:26   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Jul 26 11:05:26   at java.lang.reflect.Method.invoke(Method.java:498)
> Jul 26 11:05:26   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Jul 26 11:05:26   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Jul 26 11:05:26   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Jul 26 11:05:26   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Jul 26 11:05:26   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> Jul 26 11:05:26   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> Jul 26 11:05:26   at 
> org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45)
> Jul 26 11:05:26   at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> Jul 26 11:05:26   at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
> Jul 26 11:05:26   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
> Jul 26 11:05:26   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
> Jul 26 11:05:26   at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:413)
> Jul 26 11:05:26   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> Jul 26 11:05:26   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> Jul 26 11:05:26   

[GitHub] [flink] flinkbot edited a comment on pull request #16598: [FLINK-23501][GCP PubSub] add table source API for GCP PubSub

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16598:
URL: https://github.com/apache/flink/pull/16598#issuecomment-886655820


   
   ## CI report:
   
   * 7b7fb97cecd535742504992a9f884724bd9ed9ad Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21039)
 
   * 65f256e785b3ac499d38eee63d177f6a6059b609 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16600: [FLINK-19929] Upgrade Kinesis dependencies

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16600:
URL: https://github.com/apache/flink/pull/16600#issuecomment-886780305


   
   ## CI report:
   
   * 65a0c313d87a912a26085482d8303b316d019caa Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21025)
 
   * dccbfc290b390865ce848211f7615ef5aed35efe Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21048)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16594: [FLINK-23063][python][table] Remove #connect

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16594:
URL: https://github.com/apache/flink/pull/16594#issuecomment-886556323


   
   ## CI report:
   
   * f6870e09386e9c63459c2de0de3220ad6d028e0d UNKNOWN
   * 4b2cae270f8e378cf7ac80a2b262db2b83752023 UNKNOWN
   * 17cef7e332076311fb1af59272f22d98c8610f11 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21036)
 
   * a47808d9b556988301e767eea2731989ec3c4663 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21042)
 
   * ad84d21a38c492d65ba708c222edc295b431da29 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zentol commented on pull request #16610: [FLINK-23497][table-planner] Add dependency for scala-parser-combinators

2021-07-27 Thread GitBox


zentol commented on pull request #16610:
URL: https://github.com/apache/flink/pull/16610#issuecomment-887755621


   > do we need to do something around NOTICE files? The dependency is 
mentioned in flink-dist already but with hardcoded Scala 2.11.
   
   Is it bundled in the table-uber-jar? If so then let's add it to that NOTICE 
to avoid issues in the future.
   But if it is in flink-dist then we could also set it to provided?
   
   It's fine to not account for the scala version.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-23496) Maven snapshot build does not use MAVEN_GLOBAL_OPTIONS

2021-07-27 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-23496.

Resolution: Fixed

master: 3a9e414265d987fbdc14be84d768c63941f04d58

> Maven snapshot build does not use MAVEN_GLOBAL_OPTIONS
> --
>
> Key: FLINK-23496
> URL: https://issues.apache.org/jira/browse/FLINK-23496
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System, Build System / CI
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.14.0
>
>
> On CI we have a collection of useful settings for maven that we generally use 
> everywhere, but the deployment of maven snapshot artifacts currently doesn't.
> This leads to some duplication and noisy logs.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zentol merged pull request #16595: [FLINK-23496][ci] Use MVN_GLOBAL_OPTIONS during maven snapshot deployment

2021-07-27 Thread GitBox


zentol merged pull request #16595:
URL: https://github.com/apache/flink/pull/16595


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-23504) Rework TestingRpcService to not depend on akka

2021-07-27 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-23504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-23504.

Resolution: Fixed

master: abf9752c9de3810e10b9f7bbb937cbdf288d2789

> Rework TestingRpcService to not depend on akka
> --
>
> Key: FLINK-23504
> URL: https://issues.apache.org/jira/browse/FLINK-23504
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination, Tests
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.14.0
>
>
> The {{TestingRpcService}}, primarily used in flink-runtime, is currently an 
> extension of the AkkaRpcService. We can re-implement this class as a wrapper 
> around the {{AkkaRpcService}}.
> This would allow us to remove the test flink-rpc-akka dependencies from 
> flink-runtime.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zentol merged pull request #16601: [FLINK-23504][tests] Isolate TestingRpcService from Akka

2021-07-27 Thread GitBox


zentol merged pull request #16601:
URL: https://github.com/apache/flink/pull/16601


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] brachi-wernick commented on pull request #16598: [FLINK-23501][GCP PubSub] add table source API for GCP PubSub

2021-07-27 Thread GitBox


brachi-wernick commented on pull request #16598:
URL: https://github.com/apache/flink/pull/16598#issuecomment-887745257


   @Airblader I address most of the comment, and reply when needed.
   
   > We'll also need to update the documentation to document the PubSub 
connector for Table API.
   
   Should I just add `flink/docs/content/docs/connectors/table/pubsub.md` file? 
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16609: [FLINK-23324][connector-jdbc] fix Postgres case-insensitive.

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16609:
URL: https://github.com/apache/flink/pull/16609#issuecomment-887558020


   
   ## CI report:
   
   * 280fd415df2e9358efdfa48a675cb2a0063e878a Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21037)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16600: [FLINK-19929] Upgrade Kinesis dependencies

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16600:
URL: https://github.com/apache/flink/pull/16600#issuecomment-886780305


   
   ## CI report:
   
   * 65a0c313d87a912a26085482d8303b316d019caa Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21025)
 
   * dccbfc290b390865ce848211f7615ef5aed35efe UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16594: [FLINK-23063][python][table] Remove #connect

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16594:
URL: https://github.com/apache/flink/pull/16594#issuecomment-886556323


   
   ## CI report:
   
   * f6870e09386e9c63459c2de0de3220ad6d028e0d UNKNOWN
   * 4b2cae270f8e378cf7ac80a2b262db2b83752023 UNKNOWN
   * 17cef7e332076311fb1af59272f22d98c8610f11 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21036)
 
   * a47808d9b556988301e767eea2731989ec3c4663 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21042)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * 4d6e1fcdba2be3c29fd6bc6c1a0a4c576f91b0b6 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21035)
 
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * f2628656933e4204ee708066a9197101cfdbc72b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21044)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16611: [FLINK-22911][python] Align FLIP-136 (Improve interoperability between DataStream and Table API) in PyFlink Table API

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16611:
URL: https://github.com/apache/flink/pull/16611#issuecomment-887695133


   
   ## CI report:
   
   * 8d5954945c4eeb370bc8b3f5da3f8356b00f8278 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21047)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16556: [FLINK-23452][runtime] Added the throughput meter for measurement of …

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16556:
URL: https://github.com/apache/flink/pull/16556#issuecomment-884238774


   
   ## CI report:
   
   * 8c76db43092d6df17f52862a0d098dfb3a4a5ab3 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=20833)
 
   * 4033dd607f8ea9d375c48e3054f070221f897674 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21046)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #16589: [FLINK-23408] Wait for a checkpoint completed after finishing a task

2021-07-27 Thread GitBox


flinkbot edited a comment on pull request #16589:
URL: https://github.com/apache/flink/pull/16589#issuecomment-886429447


   
   ## CI report:
   
   * 791a3f1825b2e2d5682c40e9b7e52a4cbc576429 UNKNOWN
   * a29a6115a578c543a9856a07d0ef58f79ebcc895 UNKNOWN
   * c679fefa6bc59735ea1fb7f3ed6a5cd74685d288 UNKNOWN
   * cd0181abe1514ca324996415bca37669d879d350 UNKNOWN
   * dcd9da69c9b86992a16cf338f890fa3f9cff3cd7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21033)
 
   * 1453bb994c02b90aab31e8c5d1f4ac6ae1452e06 UNKNOWN
   * 4d6e1fcdba2be3c29fd6bc6c1a0a4c576f91b0b6 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21035)
 
   * e6574235f235d7377181c96e0390beac5d8a7ae0 UNKNOWN
   * f2628656933e4204ee708066a9197101cfdbc72b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=21044)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




  1   2   3   4   >