[GitHub] [flink] yxu-valleytider commented on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
yxu-valleytider commented on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-532995742
 
 
   PTAL @kl0u 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9714: [FLINK-14123]Change taskmanager.memory.fraction default value to 0.6

2019-09-18 Thread GitBox
flinkbot commented on issue #9714: [FLINK-14123]Change 
taskmanager.memory.fraction default value to 0.6
URL: https://github.com/apache/flink/pull/9714#issuecomment-532994026
 
 
   
   ## CI report:
   
   * 9c2f79d83a667cefc8364c4244f183e3798d9602 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9713: [hotfix][API/DataStream] Add missing assert for checkpoint lock checks

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9713: [hotfix][API/DataStream] Add missing 
assert for checkpoint lock checks
URL: https://github.com/apache/flink/pull/9713#issuecomment-532988933
 
 
   
   ## CI report:
   
   * f0fb6004eb0d0cf294f530aa0348ce23c5e701d8 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128286311)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9711: [FLINK-14033] upload user artifacts for yarn job cluster

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9711: [FLINK-14033] upload user artifacts 
for yarn job cluster
URL: https://github.com/apache/flink/pull/9711#issuecomment-532971678
 
 
   
   ## CI report:
   
   * 2011905b5abe4cb332a60bc3f70378c777482924 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128281677)
   * 8f3ec4639c0c16591302cdd2a5b294d357903a22 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128282747)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] xuyang1706 commented on a change in pull request #9184: [FLINK-13339][ml] Add an implementation of pipeline's api

2019-09-18 Thread GitBox
xuyang1706 commented on a change in pull request #9184: [FLINK-13339][ml] Add 
an implementation of pipeline's api
URL: https://github.com/apache/flink/pull/9184#discussion_r326013357
 
 

 ##
 File path: 
flink-ml-parent/flink-ml-lib/src/main/java/org/apache/flink/ml/batchoperator/BatchOperator.java
 ##
 @@ -0,0 +1,80 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
+package org.apache.flink.ml.batchoperator;
+
+import org.apache.flink.ml.api.misc.param.Params;
+import org.apache.flink.ml.batchoperator.source.TableSourceBatchOp;
+import org.apache.flink.ml.common.AlgoOperator;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.util.Preconditions;
+
+import java.util.ArrayList;
+import java.util.List;
+
+/**
+ * Base class of batch algorithm operators.
+ */
+public abstract class BatchOperator> extends 
AlgoOperator {
 
 Review comment:
   Thanks, provided.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] xuyang1706 commented on a change in pull request #9184: [FLINK-13339][ml] Add an implementation of pipeline's api

2019-09-18 Thread GitBox
xuyang1706 commented on a change in pull request #9184: [FLINK-13339][ml] Add 
an implementation of pipeline's api
URL: https://github.com/apache/flink/pull/9184#discussion_r326013260
 
 

 ##
 File path: 
flink-ml-parent/flink-ml-lib/src/main/java/org/apache/flink/ml/batchoperator/BatchOperator.java
 ##
 @@ -0,0 +1,80 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
+package org.apache.flink.ml.batchoperator;
+
+import org.apache.flink.ml.api.misc.param.Params;
+import org.apache.flink.ml.batchoperator.source.TableSourceBatchOp;
+import org.apache.flink.ml.common.AlgoOperator;
+import org.apache.flink.table.api.Table;
+import org.apache.flink.util.Preconditions;
+
+import java.util.ArrayList;
+import java.util.List;
+
+/**
+ * Base class of batch algorithm operators.
+ */
+public abstract class BatchOperator> extends 
AlgoOperator {
+
+   public BatchOperator() {
+   super();
+   }
+
+   public BatchOperator(Params params) {
+   super(params);
+   }
+
+   public > B link(B next) {
+   return linkTo(next);
+   }
+
+   public > B linkTo(B next) {
+   next.linkFrom(this);
+   return next;
+   }
+
+   public abstract T linkFrom(BatchOperator... inputs);
 
 Review comment:
   Thanks, added intro and example.
   For there is no answer area of the above question, I put the answer here:
   PipelineStage only supports single input and single output, it is the basic 
unit for pipeline. AlgoOperator supports multi-input and multi-output. We’d 
like to implement the algorithm with  AlgoOperator, and PipelineStage’s fit and 
transform function can call the AlgoOperator.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9714: [FLINK-14123]Change taskmanager.memory.fraction default value to 0.6

2019-09-18 Thread GitBox
flinkbot commented on issue #9714: [FLINK-14123]Change 
taskmanager.memory.fraction default value to 0.6
URL: https://github.com/apache/flink/pull/9714#issuecomment-532990781
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 9c2f79d83a667cefc8364c4244f183e3798d9602 (Thu Sep 19 
06:42:20 UTC 2019)
   
   **Warnings:**
* **This pull request references an unassigned [Jira 
ticket](https://issues.apache.org/jira/browse/FLINK-14123).** According to the 
[code contribution 
guide](https://flink.apache.org/contributing/contribute-code.html), tickets 
need to be assigned before starting with the implementation work.
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liupc opened a new pull request #9714: [FLINK-14123]Change taskmanager.memory.fraction default value to 0.6

2019-09-18 Thread GitBox
liupc opened a new pull request #9714: [FLINK-14123]Change 
taskmanager.memory.fraction default value to 0.6
URL: https://github.com/apache/flink/pull/9714
 
 
   
   ## What is the purpose of the change
   
   The default JVM Option value `NewRatio` is 2, that means the old generation 
ocuppies 2/3 (0.66) of the heap memory, but we allocate 0.7 
(`taskmanager.memory.fraction` default value) of the heap memory in 
`UnilateralSortMerger`, that's a little bit aggressive.
   This PR will change the default value to 0.6 to be more safe and stable.
   
   
   ## Brief change log
   
 - Changed the default value in TaskManagerOptions and docs and related 
scripts.
   
   ## Verifying this change
   
   This change is already covered by existing tests, and also tested with 
custom terasort application.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: ( no )
 - The runtime per-record code paths (performance sensitive): (no )
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no )
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-14123) Change taskmanager.memory.fraction default value to 0.6

2019-09-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-14123:
---
Labels: pull-request-available  (was: )

> Change taskmanager.memory.fraction default value to 0.6
> ---
>
> Key: FLINK-14123
> URL: https://issues.apache.org/jira/browse/FLINK-14123
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Configuration
>Affects Versions: 1.9.0
>Reporter: liupengcheng
>Priority: Major
>  Labels: pull-request-available
>
> Currently, we are testing flink batch task, such as terasort, however, it 
> started only awhile then it failed due to OOM. 
>  
> {code:java}
> org.apache.flink.client.program.ProgramInvocationException: Job failed. 
> (JobID: a807e1d635bd4471ceea4282477f8850)
>   at 
> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
>   at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
>   at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
>   at 
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:62)
>   at 
> org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:539)
>   at 
> com.github.ehiggs.spark.terasort.FlinkTeraSort$.main(FlinkTeraSort.scala:89)
>   at 
> com.github.ehiggs.spark.terasort.FlinkTeraSort.main(FlinkTeraSort.scala)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:604)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:466)
>   at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:274)
>   at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:746)
>   at 
> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:273)
>   at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:205)
>   at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1007)
>   at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1080)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
>   at 
> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>   at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1080)
> Caused by: org.apache.flink.runtime.client.JobExecutionException: Job 
> execution failed.
>   at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
>   at 
> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
>   ... 23 more
> Caused by: java.lang.RuntimeException: Error obtaining the sorted input: 
> Thread 'SortMerger Reading Thread' terminated due to an exception: GC 
> overhead limit exceeded
>   at 
> org.apache.flink.runtime.operators.sort.UnilateralSortMerger.getIterator(UnilateralSortMerger.java:650)
>   at 
> org.apache.flink.runtime.operators.BatchTask.getInput(BatchTask.java:1109)
>   at org.apache.flink.runtime.operators.NoOpDriver.run(NoOpDriver.java:82)
>   at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:504)
>   at 
> org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: java.io.IOException: Thread 'SortMerger Reading Thread' terminated 
> due to an exception: GC overhead limit exceeded
>   at 
> org.apache.flink.runtime.operators.sort.UnilateralSortMerger$ThreadBase.run(UnilateralSortMerger.java:831)
> Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded
>   at 
> org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.deserialize(BytePrimitiveArraySerializer.java:84)
>   at 
> org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.deserialize(BytePrimitiveArraySerializer.java:33)
>   at 
> org.apache.flink.api.scala.typeutils.CaseClassSerializer.deserialize(CaseClassSeriali

[GitHub] [flink] flinkbot commented on issue #9713: [hotfix][API/DataStream] Add missing assert for checkpoint lock checks

2019-09-18 Thread GitBox
flinkbot commented on issue #9713: [hotfix][API/DataStream] Add missing assert 
for checkpoint lock checks
URL: https://github.com/apache/flink/pull/9713#issuecomment-532988933
 
 
   
   ## CI report:
   
   * f0fb6004eb0d0cf294f530aa0348ce23c5e701d8 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9712: [FLINK-13656] [sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 1.21.0

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9712: [FLINK-13656] 
[sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 
1.21.0
URL: https://github.com/apache/flink/pull/9712#issuecomment-532984287
 
 
   
   ## CI report:
   
   * f2b7710f65f478342de389c8e099799287ddf3f9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128285030)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128271616)
   * 9abb55b018585e11703afe9f2494ae95fb129c0a : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128277184)
   * 0301c85818f5cd92b694dbee070cd1a99895f495 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128285009)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-14123) Change taskmanager.memory.fraction default value to 0.6

2019-09-18 Thread liupengcheng (Jira)
liupengcheng created FLINK-14123:


 Summary: Change taskmanager.memory.fraction default value to 0.6
 Key: FLINK-14123
 URL: https://issues.apache.org/jira/browse/FLINK-14123
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Configuration
Affects Versions: 1.9.0
Reporter: liupengcheng


Currently, we are testing flink batch task, such as terasort, however, it 
started only awhile then it failed due to OOM. 

 
{code:java}
org.apache.flink.client.program.ProgramInvocationException: Job failed. (JobID: 
a807e1d635bd4471ceea4282477f8850)
at 
org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
at 
org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
at 
org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
at 
org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:62)
at 
org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:539)
at 
com.github.ehiggs.spark.terasort.FlinkTeraSort$.main(FlinkTeraSort.scala:89)
at 
com.github.ehiggs.spark.terasort.FlinkTeraSort.main(FlinkTeraSort.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:604)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:466)
at 
org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:274)
at 
org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:746)
at 
org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:273)
at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:205)
at 
org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1007)
at 
org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1080)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
at 
org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1080)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution 
failed.
at 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
at 
org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
... 23 more
Caused by: java.lang.RuntimeException: Error obtaining the sorted input: Thread 
'SortMerger Reading Thread' terminated due to an exception: GC overhead limit 
exceeded
at 
org.apache.flink.runtime.operators.sort.UnilateralSortMerger.getIterator(UnilateralSortMerger.java:650)
at 
org.apache.flink.runtime.operators.BatchTask.getInput(BatchTask.java:1109)
at org.apache.flink.runtime.operators.NoOpDriver.run(NoOpDriver.java:82)
at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:504)
at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.IOException: Thread 'SortMerger Reading Thread' terminated 
due to an exception: GC overhead limit exceeded
at 
org.apache.flink.runtime.operators.sort.UnilateralSortMerger$ThreadBase.run(UnilateralSortMerger.java:831)
Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded
at 
org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.deserialize(BytePrimitiveArraySerializer.java:84)
at 
org.apache.flink.api.common.typeutils.base.array.BytePrimitiveArraySerializer.deserialize(BytePrimitiveArraySerializer.java:33)
at 
org.apache.flink.api.scala.typeutils.CaseClassSerializer.deserialize(CaseClassSerializer.scala:121)
at 
org.apache.flink.api.scala.typeutils.CaseClassSerializer.deserialize(CaseClassSerializer.scala:114)
at 
org.apache.flink.api.scala.typeutils.CaseClassSerializer.deserialize(CaseClassSerializer.scala:32)
at 
org.apache.flink.runtime.plugable.ReusingDeserializationDelegate.read(ReusingDeserializationDelegate.java:57)
at 
org.ap

[jira] [Comment Edited] (FLINK-13682) Translate "Code Style - Scala Guide" page into Chinese

2019-09-18 Thread Jeff Yang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13682?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933091#comment-16933091
 ] 

Jeff Yang edited comment on FLINK-13682 at 9/19/19 6:23 AM:


Hi, [~jark]  I am already done ,Please  take a look 
[PR|https://github.com/apache/flink-web/pull/267],Thanks .


was (Author: highfei2...@126.com):
Hi, [~jark]  I am already done ,Please  take a look this 
[PR|[https://github.com/apache/flink-web/pull/267]],Thanks .

> Translate "Code Style - Scala Guide" page into Chinese
> --
>
> Key: FLINK-13682
> URL: https://issues.apache.org/jira/browse/FLINK-13682
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: Jeff Yang
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Translate page 
> https://flink.apache.org/zh/contributing/code-style-and-quality-scala.html 
> into Chinese. The page is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/code-style-and-quality-scala.zh.md.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #9713: [hotfix][API/DataStream] Add missing assert for checkpoint lock checks

2019-09-18 Thread GitBox
flinkbot commented on issue #9713: [hotfix][API/DataStream] Add missing assert 
for checkpoint lock checks
URL: https://github.com/apache/flink/pull/9713#issuecomment-532985575
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit f0fb6004eb0d0cf294f530aa0348ce23c5e701d8 (Thu Sep 19 
06:23:41 UTC 2019)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13682) Translate "Code Style - Scala Guide" page into Chinese

2019-09-18 Thread Jeff Yang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13682?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933091#comment-16933091
 ] 

Jeff Yang commented on FLINK-13682:
---

Hi, [~jark]  I am already done ,Please  take a look this 
[PR|[https://github.com/apache/flink-web/pull/267]],Thanks .

> Translate "Code Style - Scala Guide" page into Chinese
> --
>
> Key: FLINK-13682
> URL: https://issues.apache.org/jira/browse/FLINK-13682
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: Jeff Yang
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Translate page 
> https://flink.apache.org/zh/contributing/code-style-and-quality-scala.html 
> into Chinese. The page is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/code-style-and-quality-scala.zh.md.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] link3280 opened a new pull request #9713: [hotfix][API/DataStream] Add missing assert for checkpoint lock checks

2019-09-18 Thread GitBox
link3280 opened a new pull request #9713: [hotfix][API/DataStream] Add missing 
assert for checkpoint lock checks
URL: https://github.com/apache/flink/pull/9713
 
 
   ## What is the purpose of the change
   
   The return value of `Thread.holdsLock(checkpointLock)` is unused, which are 
supposed to be validated by an assertion instead.
   
   ## Brief change log
   
   - Add assert clause to validate the return value of 
`Thread.holdsLock(checkpointLock)`.
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: ( no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13682) Translate "Code Style - Scala Guide" page into Chinese

2019-09-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13682?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-13682:
---
Labels: pull-request-available  (was: )

> Translate "Code Style - Scala Guide" page into Chinese
> --
>
> Key: FLINK-13682
> URL: https://issues.apache.org/jira/browse/FLINK-13682
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: Jeff Yang
>Priority: Major
>  Labels: pull-request-available
>
> Translate page 
> https://flink.apache.org/zh/contributing/code-style-and-quality-scala.html 
> into Chinese. The page is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/code-style-and-quality-scala.zh.md.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #9712: [FLINK-13656] [sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 1.21.0

2019-09-18 Thread GitBox
flinkbot commented on issue #9712: [FLINK-13656] 
[sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 
1.21.0
URL: https://github.com/apache/flink/pull/9712#issuecomment-532984287
 
 
   
   ## CI report:
   
   * f2b7710f65f478342de389c8e099799287ddf3f9 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9683: [FLINK-14037]Use same userCodeClassLoader as jobManager in jobClient to fix input/…

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9683: [FLINK-14037]Use same 
userCodeClassLoader as jobManager in jobClient to fix input/…
URL: https://github.com/apache/flink/pull/9683#issuecomment-531697778
 
 
   
   ## CI report:
   
   * ba87beba470b132b99ed606411ffc43847a72d8f : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/127766962)
   * 9838e772ccac5488aee7fb52837f21bcce94d882 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/127936152)
   * 89765aad20b64adb2c240f818af45037f6e36e0d : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128133420)
   * 421a64587f14f5116f6375f7b285bc6846bdc6ff : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128283861)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128271616)
   * 9abb55b018585e11703afe9f2494ae95fb129c0a : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128277184)
   * 0301c85818f5cd92b694dbee070cd1a99895f495 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] yxu-valleytider commented on a change in pull request #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileS

2019-09-18 Thread GitBox
yxu-valleytider commented on a change in pull request #9581: 
[FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to 
allow for easier subclassing of StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#discussion_r326002753
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/filesystem/StreamingFileSink.java
 ##
 @@ -162,47 +162,54 @@ protected StreamingFileSink(
 * @return The builder where the remaining of the configuration 
parameters for the sink can be configured.
 * In order to instantiate the sink, call {@link 
RowFormatBuilder#build()} after specifying the desired parameters.
 */
-   public static  StreamingFileSink.BulkFormatBuilder 
forBulkFormat(
+   public static  StreamingFileSink.BulkFormatBuilder> forBulkFormat(
final Path basePath, final BulkWriter.Factory 
writerFactory) {
return new StreamingFileSink.BulkFormatBuilder<>(basePath, 
writerFactory, new DateTimeBucketAssigner<>());
}
 
/**
 * The base abstract class for the {@link RowFormatBuilder} and {@link 
BulkFormatBuilder}.
 */
-   protected abstract static class BucketsBuilder implements 
Serializable {
+   protected abstract static class BucketsBuilder> implements Serializable {
 
 Review comment:
   DONE.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] yxu-valleytider commented on a change in pull request #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileS

2019-09-18 Thread GitBox
yxu-valleytider commented on a change in pull request #9581: 
[FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to 
allow for easier subclassing of StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#discussion_r326000120
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/filesystem/StreamingFileSink.java
 ##
 @@ -264,37 +288,32 @@ private RowFormatBuilder(
subtaskIndex,
new PartFileConfig(partFilePrefix, 
partFileSuffix));
}
-
-   @VisibleForTesting
-   StreamingFileSink.RowFormatBuilder 
withBucketFactory(final BucketFactory factory) {
-   return new RowFormatBuilder<>(basePath, encoder, 
bucketAssigner, rollingPolicy, bucketCheckInterval, 
Preconditions.checkNotNull(factory), partFilePrefix, partFileSuffix);
-   }
}
 
/**
 * A builder for configuring the sink for bulk-encoding formats, e.g. 
Parquet/ORC.
 */
@PublicEvolving
-   public static class BulkFormatBuilder extends 
StreamingFileSink.BucketsBuilder {
+   public static class BulkFormatBuilder> extends 
StreamingFileSink.BucketsBuilder {
 
 Review comment:
   DONE similar changes as `RowFormatBuilder`. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14037) Deserializing the input/output formats failed: unread block data

2019-09-18 Thread liupengcheng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14037?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933087#comment-16933087
 ] 

liupengcheng commented on FLINK-14037:
--

[~zhuzh] 

Yes, that's weird, Maybe I think your test is not complete, you can change 
x.y.z.B to a.b.c.B, and test it again to check whether the results are still 
the same. I'm not sure about the behavior of the defineClass, but I think the 
delegation loading only happen in `loadClass` method, in defineClass, it will 
load with current loader. Maybe I'm wrong, but if we cannot explain my case, 
then I'm not convinced that `alwaysParentFirstLoaderPatterns` will always make 
class in the list loaded by parent loader unless it does not exist in the 
parent classloader. .

 

> Deserializing the input/output formats failed: unread block data
> 
>
> Key: FLINK-14037
> URL: https://issues.apache.org/jira/browse/FLINK-14037
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.9.0
> Environment: flink 1.9.0
> app jar use `flink-shaded-hadoop-2` dependencies to avoid some confilicts
>  
>Reporter: liupengcheng
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Recently, we encountered the following issue when testing flink 1.9.0:
> {code:java}
> org.apache.flink.client.program.ProgramInvocationException: Could not 
> retrieve the execution result. (JobID: 8ffbc071dda81d6f8005c02be8adde6b)
>   at 
> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:255)
>   at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
>   at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:326)
>   at 
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:62)
>   at 
> org.apache.flink.api.scala.ExecutionEnvironment.execute(ExecutionEnvironment.scala:539)
>   at 
> com.github.ehiggs.spark.terasort.FlinkTeraSort$.main(FlinkTeraSort.scala:89)
>   at 
> com.github.ehiggs.spark.terasort.FlinkTeraSort.main(FlinkTeraSort.scala)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:576)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:438)
>   at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:274)
>   at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:746)
>   at 
> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:273)
>   at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:205)
>   at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1010)
>   at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1083)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
>   at 
> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>   at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1083)
> Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed to 
> submit JobGraph.
>   at 
> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$8(RestClusterClient.java:382)
>   at 
> java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:870)
>   at 
> java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:852)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>   at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>   at 
> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:263)
>   at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>   at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>   at 
> java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:5

[GitHub] [flink] yxu-valleytider commented on a change in pull request #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileS

2019-09-18 Thread GitBox
yxu-valleytider commented on a change in pull request #9581: 
[FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to 
allow for easier subclassing of StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#discussion_r325999744
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/filesystem/StreamingFileSink.java
 ##
 @@ -224,35 +231,52 @@ private RowFormatBuilder(
this.partFileSuffix = 
Preconditions.checkNotNull(partFileSuffix);
}
 
-   public StreamingFileSink.RowFormatBuilder 
withBucketCheckInterval(final long interval) {
-   return new RowFormatBuilder<>(basePath, encoder, 
bucketAssigner, rollingPolicy, interval, bucketFactory, partFilePrefix, 
partFileSuffix);
+   public T withBucketCheckInterval(final long interval) {
+   this.bucketCheckInterval = interval;
+   return self();
}
 
-   public StreamingFileSink.RowFormatBuilder 
withBucketAssigner(final BucketAssigner assigner) {
-   return new RowFormatBuilder<>(basePath, encoder, 
Preconditions.checkNotNull(assigner), rollingPolicy, bucketCheckInterval, 
bucketFactory, partFilePrefix, partFileSuffix);
+   public T withBucketAssigner(final BucketAssigner 
assigner) {
+   this.bucketAssigner = 
Preconditions.checkNotNull(assigner);
+   return self();
}
 
-   public StreamingFileSink.RowFormatBuilder 
withRollingPolicy(final RollingPolicy policy) {
-   return new RowFormatBuilder<>(basePath, encoder, 
bucketAssigner, Preconditions.checkNotNull(policy), bucketCheckInterval, 
bucketFactory, partFilePrefix, partFileSuffix);
+   public T withRollingPolicy(final RollingPolicy 
policy) {
+   this.rollingPolicy = Preconditions.checkNotNull(policy);
+   return self();
}
 
-   public  StreamingFileSink.RowFormatBuilder 
withBucketAssignerAndPolicy(final BucketAssigner assigner, final 
RollingPolicy policy) {
-   return new RowFormatBuilder<>(basePath, encoder, 
Preconditions.checkNotNull(assigner), Preconditions.checkNotNull(policy), 
bucketCheckInterval, new DefaultBucketFactoryImpl<>(), partFilePrefix, 
partFileSuffix);
+   public T withBucketAssignerAndPolicy(final BucketAssigner assigner, final RollingPolicy policy) {
+   this.bucketAssigner = 
Preconditions.checkNotNull(assigner);
+   this.rollingPolicy = Preconditions.checkNotNull(policy);
+   return self();
}
 
-   public StreamingFileSink.RowFormatBuilder 
withPartFilePrefix(final String partPrefix) {
-   return new RowFormatBuilder<>(basePath, encoder, 
bucketAssigner, rollingPolicy, bucketCheckInterval, bucketFactory, partPrefix, 
partFileSuffix);
+   public T withPartFilePrefix(final String partPrefix) {
+   this.partFilePrefix = partPrefix;
+   return self();
}
 
-   public StreamingFileSink.RowFormatBuilder 
withPartFileSuffix(final String partSuffix) {
-   return new RowFormatBuilder<>(basePath, encoder, 
bucketAssigner, rollingPolicy, bucketCheckInterval, bucketFactory, 
partFilePrefix, partSuffix);
+   public T withPartFileSuffix(final String partSuffix) {
+   this.partFileSuffix = partSuffix;
+   return self();
+   }
+
+   public  StreamingFileSink.RowFormatBuilder> 
newBuilderWithBucketAssignerAndPolicy(final BucketAssigner assigner, 
final RollingPolicy policy) {
+   return new RowFormatBuilder<>(basePath, encoder, 
Preconditions.checkNotNull(assigner), Preconditions.checkNotNull(policy), 
bucketCheckInterval, new DefaultBucketFactoryImpl<>(), partFilePrefix, 
partFileSuffix);
 
 Review comment:
   DONE.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] yxu-valleytider commented on a change in pull request #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileS

2019-09-18 Thread GitBox
yxu-valleytider commented on a change in pull request #9581: 
[FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to 
allow for easier subclassing of StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#discussion_r326002736
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/filesystem/StreamingFileSink.java
 ##
 @@ -162,47 +162,54 @@ protected StreamingFileSink(
 * @return The builder where the remaining of the configuration 
parameters for the sink can be configured.
 * In order to instantiate the sink, call {@link 
RowFormatBuilder#build()} after specifying the desired parameters.
 */
-   public static  StreamingFileSink.BulkFormatBuilder 
forBulkFormat(
+   public static  StreamingFileSink.BulkFormatBuilder> forBulkFormat(
final Path basePath, final BulkWriter.Factory 
writerFactory) {
return new StreamingFileSink.BulkFormatBuilder<>(basePath, 
writerFactory, new DateTimeBucketAssigner<>());
}
 
/**
 * The base abstract class for the {@link RowFormatBuilder} and {@link 
BulkFormatBuilder}.
 */
-   protected abstract static class BucketsBuilder implements 
Serializable {
+   protected abstract static class BucketsBuilder> implements Serializable {
 
private static final long serialVersionUID = 1L;
 
+   protected static final long DEFAULT_BUCKET_CHECK_INTERVAL = 60L 
* 1000L;
+
+   @SuppressWarnings("unchecked")
+   protected T self() {
+   return (T) this;
+   }
+
abstract Buckets createBuckets(final int 
subtaskIndex) throws IOException;
}
 
/**
 * A builder for configuring the sink for row-wise encoding formats.
 */
@PublicEvolving
-   public static class RowFormatBuilder extends 
StreamingFileSink.BucketsBuilder {
+   public static class RowFormatBuilder> extends StreamingFileSink.BucketsBuilder {
 
private static final long serialVersionUID = 1L;
 
-   private final long bucketCheckInterval;
+   protected long bucketCheckInterval;
 
 Review comment:
   DONE.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] yxu-valleytider commented on a change in pull request #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileS

2019-09-18 Thread GitBox
yxu-valleytider commented on a change in pull request #9581: 
[FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to 
allow for easier subclassing of StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#discussion_r326002700
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/filesystem/StreamingFileSink.java
 ##
 @@ -162,47 +162,54 @@ protected StreamingFileSink(
 * @return The builder where the remaining of the configuration 
parameters for the sink can be configured.
 * In order to instantiate the sink, call {@link 
RowFormatBuilder#build()} after specifying the desired parameters.
 */
-   public static  StreamingFileSink.BulkFormatBuilder 
forBulkFormat(
+   public static  StreamingFileSink.BulkFormatBuilder> forBulkFormat(
final Path basePath, final BulkWriter.Factory 
writerFactory) {
return new StreamingFileSink.BulkFormatBuilder<>(basePath, 
writerFactory, new DateTimeBucketAssigner<>());
}
 
/**
 * The base abstract class for the {@link RowFormatBuilder} and {@link 
BulkFormatBuilder}.
 */
-   protected abstract static class BucketsBuilder implements 
Serializable {
+   protected abstract static class BucketsBuilder> implements Serializable {
 
private static final long serialVersionUID = 1L;
 
+   protected static final long DEFAULT_BUCKET_CHECK_INTERVAL = 60L 
* 1000L;
+
+   @SuppressWarnings("unchecked")
+   protected T self() {
+   return (T) this;
+   }
+
abstract Buckets createBuckets(final int 
subtaskIndex) throws IOException;
}
 
/**
 * A builder for configuring the sink for row-wise encoding formats.
 */
@PublicEvolving
-   public static class RowFormatBuilder extends 
StreamingFileSink.BucketsBuilder {
+   public static class RowFormatBuilder> extends StreamingFileSink.BucketsBuilder {
 
private static final long serialVersionUID = 1L;
 
-   private final long bucketCheckInterval;
+   protected long bucketCheckInterval;
 
private final Path basePath;
 
-   private final Encoder encoder;
+   private Encoder encoder;
 
-   private final BucketAssigner bucketAssigner;
+   private BucketAssigner bucketAssigner;
 
-   private final RollingPolicy rollingPolicy;
+   private RollingPolicy rollingPolicy;
 
-   private final BucketFactory bucketFactory;
+   private BucketFactory bucketFactory;
 
-   private final String partFilePrefix;
+   private String partFilePrefix;
 
-   private final String partFileSuffix;
+   private String partFileSuffix;
 
-   RowFormatBuilder(Path basePath, Encoder encoder, 
BucketAssigner bucketAssigner) {
-   this(basePath, encoder, bucketAssigner, 
DefaultRollingPolicy.builder().build(), 60L * 1000L, new 
DefaultBucketFactoryImpl<>(), PartFileConfig.DEFAULT_PART_PREFIX, 
PartFileConfig.DEFAULT_PART_SUFFIX);
+   protected RowFormatBuilder(Path basePath, Encoder encoder, 
BucketAssigner bucketAssigner) {
+   this(basePath, encoder, bucketAssigner, 
DefaultRollingPolicy.builder().build(), DEFAULT_BUCKET_CHECK_INTERVAL, new 
DefaultBucketFactoryImpl<>(), PartFileConfig.DEFAULT_PART_PREFIX, 
PartFileConfig.DEFAULT_PART_SUFFIX);
}
 
private RowFormatBuilder(
 
 Review comment:
   DONE.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9711: [FLINK-14033] upload user artifacts for yarn job cluster

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9711: [FLINK-14033] upload user artifacts 
for yarn job cluster
URL: https://github.com/apache/flink/pull/9711#issuecomment-532971678
 
 
   
   ## CI report:
   
   * 2011905b5abe4cb332a60bc3f70378c777482924 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128281677)
   * 8f3ec4639c0c16591302cdd2a5b294d357903a22 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128282747)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9683: [FLINK-14037]Use same userCodeClassLoader as jobManager in jobClient to fix input/…

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9683: [FLINK-14037]Use same 
userCodeClassLoader as jobManager in jobClient to fix input/…
URL: https://github.com/apache/flink/pull/9683#issuecomment-531697778
 
 
   
   ## CI report:
   
   * ba87beba470b132b99ed606411ffc43847a72d8f : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/127766962)
   * 9838e772ccac5488aee7fb52837f21bcce94d882 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/127936152)
   * 89765aad20b64adb2c240f818af45037f6e36e0d : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128133420)
   * 421a64587f14f5116f6375f7b285bc6846bdc6ff : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9712: [FLINK-13656] [sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 1.21.0

2019-09-18 Thread GitBox
flinkbot commented on issue #9712: [FLINK-13656] 
[sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 
1.21.0
URL: https://github.com/apache/flink/pull/9712#issuecomment-532978281
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit f2b7710f65f478342de389c8e099799287ddf3f9 (Thu Sep 19 
05:54:17 UTC 2019)
   
   **Warnings:**
* **4 pom.xml files were touched**: Check for build and licensing issues.
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
* **This pull request references an unassigned [Jira 
ticket](https://issues.apache.org/jira/browse/FLINK-13656).** According to the 
[code contribution 
guide](https://flink.apache.org/contributing/contribute-code.html), tickets 
need to be assigned before starting with the implementation work.
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13656) Upgrade Calcite dependency to 1.21

2019-09-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13656?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-13656:
---
Labels: pull-request-available  (was: )

> Upgrade Calcite dependency to 1.21
> --
>
> Key: FLINK-13656
> URL: https://issues.apache.org/jira/browse/FLINK-13656
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Legacy Planner, Table SQL / Planner
>Reporter: Jark Wu
>Priority: Major
>  Labels: pull-request-available
>
> Umbrella issue for all tasks related to the next Calcite upgrade to 1.21.x 
> release
> Calcite 1.21 has been released recently, we need to upgrade to version 1.21 
> for these reasons:
> - Previously we have made some temp code to support full data types in sql 
> parser, since CALCITE-3213 has been resolved, we can do some refactoring for 
> these codes;
> - We also fixed some important bug for Join which bring in from Calcite 1.20 
> join like expression promotion, such as CALCITE-3170, CALCITE-3171.
> - CALCITE-2302 has been resolved, there is possibility we support implicit 
> type coercion for Flink now.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] liupc commented on issue #9683: [FLINK-14037]Use same userCodeClassLoader as jobManager in jobClient to fix input/…

2019-09-18 Thread GitBox
liupc commented on issue #9683: [FLINK-14037]Use same userCodeClassLoader as 
jobManager in jobClient to fix input/…
URL: https://github.com/apache/flink/pull/9683#issuecomment-532977900
 
 
   @aljoscha @TisonKun I find it's hard to make the classloader outside the 
packagedProgram, there are lots of logic and deps in PackagedProgram to build 
the classpaths for the classsloader. So I just get the 
`classLoaderResoloverOrder` and `alwaysParentFirstLoaderPatterns` at client and 
pass them to the constructor arguments of the PackagedProgram.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] danny0405 opened a new pull request #9712: [FLINK-13656] [sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 1.21.0

2019-09-18 Thread GitBox
danny0405 opened a new pull request #9712: [FLINK-13656] 
[sql-parser][table-planner][table-planner-blink] Bump Calcite dependency to 
1.21.0
URL: https://github.com/apache/flink/pull/9712
 
 
   ### SQL Parser
   Some refactoring in flink-sql-parser
   
   ### Flink Planner
   - `org.apache.calcite.rel.core.Match#partitionKeys` has changed from 
`List` to `ImmutableBitSet`, since CALCITE-1935,
   - `org.apache.calcite.rel.core.Match#copy` has been removed since 
CALCITE-1935 part1.
   - 
`org.apache.flink.table.api.stream.sql.JoinTest#testRightOuterJoinEquiAndLocalPred`,
 `#testLeftOuterJoinEquiAndLocalPred`, 
`org.apache.flink.table.api.batch.sql.JoinTest#testFullOuterJoinEquiAndLocalPred`,
 `#testRightOuterJoinEquiAndLocalPred`, `#testLeftOuterJoinEquiAndLocalPred` 
plan has changed, but it is still correct. since CALCITE-3101
   - ConverterRule that use `RelOptRule#description` was affected because the
   default description generation rule has changed since CALCITE-3115.
   - 
`org.apache.flink.table.api.stream.table.AggregateTest#testMultiDistinctAggregateSameFieldOnHopWindow`,
 
`org.apache.flink.table.api.stream.sql.DistinctAggregateTest#testMultiDistinctAggregateSameFieldOnHopWindow`
 max distinct was removed. Since CALCITE-3159
   
   ### Blink Planner
   
   - The plan change mostly because of the same reason with flink planner
   - `SqlFunction` was copied from CALCITE because of CALCITE-3360
   - org.apache.flink.table.planner.plan.batch.sql#testMinusWithNestedTypes IS 
NOT DISTINCT FROM is not expanded now since CALCITE-3174
   - `FlinkTypeFactory#createDecimalQuotient` was moved to 
`FlinkTypeSystem#deriveDecimalDivideType `due to CALCITE-3187
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9710: [FLINK-11859][runtime]Improve SpanningRecordSerializer performance by serializing record length to data buffer directly.

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9710: [FLINK-11859][runtime]Improve 
SpanningRecordSerializer performance by serializing record length to data 
buffer directly.
URL: https://github.com/apache/flink/pull/9710#issuecomment-532961713
 
 
   
   ## CI report:
   
   * 9df06c32a009c53dce3f13c7de12892eb8563213 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128278955)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9711: [FLINK-14033] upload user artifacts for yarn job cluster

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9711: [FLINK-14033] upload user artifacts 
for yarn job cluster
URL: https://github.com/apache/flink/pull/9711#issuecomment-532971678
 
 
   
   ## CI report:
   
   * 2011905b5abe4cb332a60bc3f70378c777482924 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128281677)
   * 8f3ec4639c0c16591302cdd2a5b294d357903a22 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9711: [FLINK-14033] upload user artifacts for yarn job cluster

2019-09-18 Thread GitBox
flinkbot commented on issue #9711: [FLINK-14033] upload user artifacts for yarn 
job cluster
URL: https://github.com/apache/flink/pull/9711#issuecomment-532971678
 
 
   
   ## CI report:
   
   * 2011905b5abe4cb332a60bc3f70378c777482924 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14033) Distributed caches are not registered in Yarn Per Job Cluster Mode

2019-09-18 Thread Zhenqiu Huang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14033?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933057#comment-16933057
 ] 

Zhenqiu Huang commented on FLINK-14033:
---

[~trohrmann][~fly_in_gis]
Would you please take a look at the fix? Thank you in advance.

> Distributed caches are not registered in Yarn Per Job Cluster Mode
> --
>
> Key: FLINK-14033
> URL: https://issues.apache.org/jira/browse/FLINK-14033
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.9.0
>Reporter: Zhenqiu Huang
>Assignee: Zhenqiu Huang
>Priority: Minor
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> CacheFiles in StreamExecutionEnvironment is not used in Job Submission in the 
> Yarn per job cluster mode. Compare to the job submission in session cluster 
> mode that will upload distributed cache files onto http server in application 
> master, we should get the cache files in job graph and register them into 
> blob store in YarnJobClusterEntrypoint.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #9711: [FLINK-14033] upload user artifacts for yarn job cluster

2019-09-18 Thread GitBox
flinkbot commented on issue #9711: [FLINK-14033] upload user artifacts for yarn 
job cluster
URL: https://github.com/apache/flink/pull/9711#issuecomment-532968944
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 2011905b5abe4cb332a60bc3f70378c777482924 (Thu Sep 19 
05:13:41 UTC 2019)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-14033) Distributed caches are not registered in Yarn Per Job Cluster Mode

2019-09-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-14033:
---
Labels: pull-request-available  (was: )

> Distributed caches are not registered in Yarn Per Job Cluster Mode
> --
>
> Key: FLINK-14033
> URL: https://issues.apache.org/jira/browse/FLINK-14033
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.9.0
>Reporter: Zhenqiu Huang
>Assignee: Zhenqiu Huang
>Priority: Minor
>  Labels: pull-request-available
>
> CacheFiles in StreamExecutionEnvironment is not used in Job Submission in the 
> Yarn per job cluster mode. Compare to the job submission in session cluster 
> mode that will upload distributed cache files onto http server in application 
> master, we should get the cache files in job graph and register them into 
> blob store in YarnJobClusterEntrypoint.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] HuangZhenQiu opened a new pull request #9711: [FLINK-14033] upload user artifacts for yarn job cluster

2019-09-18 Thread GitBox
HuangZhenQiu opened a new pull request #9711: [FLINK-14033] upload user 
artifacts for yarn job cluster
URL: https://github.com/apache/flink/pull/9711
 
 
   ## What is the purpose of the change
   
   Upload user artifacts that registered in distributed caches for yarn job 
clusters. Currently, distributed caches are not handled correctly. The solution 
proposed is to upload local files onto the remote file system and update the 
registered file in cache entry. 
   
   ## Brief change-log
 - Modified the AbstractClusterDescriptor to handle with artifacts for per 
job mode before starting 
application master.
   
   ## Verifying this change
   
   *(Please pick either of the following options)*
   This change is a trivial rework without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] carp84 commented on a change in pull request #9501: [FLINK-12697] [State Backends] Support on-disk state storage for spill-able heap backend

2019-09-18 Thread GitBox
carp84 commented on a change in pull request #9501: [FLINK-12697] [State 
Backends] Support on-disk state storage for spill-able heap backend
URL: https://github.com/apache/flink/pull/9501#discussion_r325990356
 
 

 ##
 File path: 
flink-state-backends/flink-statebackend-heap-spillable/src/main/java/org/apache/flink/runtime/state/heap/CopyOnWriteSkipListStateMap.java
 ##
 @@ -0,0 +1,1527 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.state.heap;
+
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.core.memory.ByteBufferUtils;
+import org.apache.flink.runtime.state.StateEntry;
+import org.apache.flink.runtime.state.StateTransformationFunction;
+import org.apache.flink.runtime.state.heap.space.Allocator;
+import org.apache.flink.runtime.state.heap.space.Chunk;
+import org.apache.flink.runtime.state.heap.space.SpaceUtils;
+import org.apache.flink.runtime.state.internal.InternalKvState;
+import org.apache.flink.util.Preconditions;
+import org.apache.flink.util.ResourceGuard;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import javax.annotation.Nonnull;
+
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Collections;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.NoSuchElementException;
+import java.util.Set;
+import java.util.Spliterators;
+import java.util.TreeSet;
+import java.util.concurrent.ConcurrentHashMap;
+import java.util.concurrent.ThreadLocalRandom;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.stream.Stream;
+import java.util.stream.StreamSupport;
+
+import static org.apache.flink.runtime.state.heap.SkipListUtils.HEAD_NODE;
+import static org.apache.flink.runtime.state.heap.SkipListUtils.NIL_NODE;
+import static 
org.apache.flink.runtime.state.heap.SkipListUtils.NIL_VALUE_POINTER;
+
+/**
+ * Implementation of state map which is based on skip list with copy-on-write 
support. states will
+ * be serialized to bytes and stored in the space allocated with the given 
allocator.
+ *
+ * @param  type of key
+ * @param  type of namespace
+ * @param  type of state
+ */
+public class CopyOnWriteSkipListStateMap extends StateMap 
implements AutoCloseable {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(CopyOnWriteSkipListStateMap.class);
+
+   /**
+* Default max number of logically-removed keys to delete one time.
+*/
+   private static final int DEFAULT_MAX_KEYS_TO_DELETE_ONE_TIME = 3;
+
+   /**
+* Default ratio of the logically-removed keys to trigger deletion when 
snapshot.
+*/
+   private static final float DEFAULT_LOGICAL_REMOVED_KEYS_RATIO = 0.2f;
+
+   /**
+* The serializer used to serialize the key and namespace to bytes 
stored in skip list.
+*/
+   private final SkipListKeySerializer skipListKeySerializer;
+
+   /**
+* The serializer used to serialize the state to bytes stored in skip 
list.
+*/
+   private final SkipListValueSerializer skipListValueSerializer;
+
+   /**
+* Space allocator.
+*/
+   private final Allocator spaceAllocator;
+
+   /**
+* The level index header.
+*/
+   private final LevelIndexHeader levelIndexHeader;
+
+   /**
+* Seed to generate random index level.
+*/
+   private int randomSeed;
+
+   /**
+* The current version of this map. Used for copy-on-write mechanics.
+*/
+   private int stateMapVersion;
+
+   /**
+* The highest version of this map that is still required by any 
unreleased snapshot.
+*/
+   private int highestRequiredSnapshotVersion;
+
+   /**
+* Snapshots no more than this version must have been finished, but 
there may be some
+* snapshots more than this version are still running.
+*/
+   private volatile int highestFinishedSnapshotVersion;
+
+   /**
+* Maintains an ordered set of version ids that are still us

[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128271616)
   * 9abb55b018585e11703afe9f2494ae95fb129c0a : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/128277184)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9710: [FLINK-11859][runtime]Improve SpanningRecordSerializer performance by serializing record length to data buffer directly.

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9710: [FLINK-11859][runtime]Improve 
SpanningRecordSerializer performance by serializing record length to data 
buffer directly.
URL: https://github.com/apache/flink/pull/9710#issuecomment-532961713
 
 
   
   ## CI report:
   
   * 9df06c32a009c53dce3f13c7de12892eb8563213 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128278955)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] carp84 commented on a change in pull request #9501: [FLINK-12697] [State Backends] Support on-disk state storage for spill-able heap backend

2019-09-18 Thread GitBox
carp84 commented on a change in pull request #9501: [FLINK-12697] [State 
Backends] Support on-disk state storage for spill-able heap backend
URL: https://github.com/apache/flink/pull/9501#discussion_r325987134
 
 

 ##
 File path: 
flink-state-backends/flink-statebackend-heap-spillable/src/main/java/org/apache/flink/runtime/state/heap/CopyOnWriteSkipListStateMap.java
 ##
 @@ -0,0 +1,1527 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.state.heap;
+
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.core.memory.ByteBufferUtils;
+import org.apache.flink.runtime.state.StateEntry;
+import org.apache.flink.runtime.state.StateTransformationFunction;
+import org.apache.flink.runtime.state.heap.space.Allocator;
+import org.apache.flink.runtime.state.heap.space.Chunk;
+import org.apache.flink.runtime.state.heap.space.SpaceUtils;
+import org.apache.flink.runtime.state.internal.InternalKvState;
+import org.apache.flink.util.Preconditions;
+import org.apache.flink.util.ResourceGuard;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import javax.annotation.Nonnull;
+
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Collections;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.NoSuchElementException;
+import java.util.Set;
+import java.util.Spliterators;
+import java.util.TreeSet;
+import java.util.concurrent.ConcurrentHashMap;
+import java.util.concurrent.ThreadLocalRandom;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.stream.Stream;
+import java.util.stream.StreamSupport;
+
+import static org.apache.flink.runtime.state.heap.SkipListUtils.HEAD_NODE;
+import static org.apache.flink.runtime.state.heap.SkipListUtils.NIL_NODE;
+import static 
org.apache.flink.runtime.state.heap.SkipListUtils.NIL_VALUE_POINTER;
+
+/**
+ * Implementation of state map which is based on skip list with copy-on-write 
support. states will
+ * be serialized to bytes and stored in the space allocated with the given 
allocator.
+ *
+ * @param  type of key
+ * @param  type of namespace
+ * @param  type of state
+ */
+public class CopyOnWriteSkipListStateMap extends StateMap 
implements AutoCloseable {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(CopyOnWriteSkipListStateMap.class);
+
+   /**
+* Default max number of logically-removed keys to delete one time.
+*/
+   private static final int DEFAULT_MAX_KEYS_TO_DELETE_ONE_TIME = 3;
+
+   /**
+* Default ratio of the logically-removed keys to trigger deletion when 
snapshot.
+*/
+   private static final float DEFAULT_LOGICAL_REMOVED_KEYS_RATIO = 0.2f;
+
+   /**
+* The serializer used to serialize the key and namespace to bytes 
stored in skip list.
+*/
+   private final SkipListKeySerializer skipListKeySerializer;
+
+   /**
+* The serializer used to serialize the state to bytes stored in skip 
list.
+*/
+   private final SkipListValueSerializer skipListValueSerializer;
+
+   /**
+* Space allocator.
+*/
+   private final Allocator spaceAllocator;
+
+   /**
+* The level index header.
+*/
+   private final LevelIndexHeader levelIndexHeader;
+
+   /**
+* Seed to generate random index level.
+*/
+   private int randomSeed;
+
+   /**
+* The current version of this map. Used for copy-on-write mechanics.
+*/
+   private int stateMapVersion;
+
+   /**
+* The highest version of this map that is still required by any 
unreleased snapshot.
+*/
+   private int highestRequiredSnapshotVersion;
+
+   /**
+* Snapshots no more than this version must have been finished, but 
there may be some
+* snapshots more than this version are still running.
+*/
+   private volatile int highestFinishedSnapshotVersion;
+
+   /**
+* Maintains an ordered set of version ids that are still us

[GitHub] [flink] carp84 commented on a change in pull request #9501: [FLINK-12697] [State Backends] Support on-disk state storage for spill-able heap backend

2019-09-18 Thread GitBox
carp84 commented on a change in pull request #9501: [FLINK-12697] [State 
Backends] Support on-disk state storage for spill-able heap backend
URL: https://github.com/apache/flink/pull/9501#discussion_r325986596
 
 

 ##
 File path: 
flink-state-backends/flink-statebackend-heap-spillable/src/main/java/org/apache/flink/runtime/state/heap/CopyOnWriteSkipListStateMap.java
 ##
 @@ -0,0 +1,1527 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.state.heap;
+
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.core.memory.ByteBufferUtils;
+import org.apache.flink.runtime.state.StateEntry;
+import org.apache.flink.runtime.state.StateTransformationFunction;
+import org.apache.flink.runtime.state.heap.space.Allocator;
+import org.apache.flink.runtime.state.heap.space.Chunk;
+import org.apache.flink.runtime.state.heap.space.SpaceUtils;
+import org.apache.flink.runtime.state.internal.InternalKvState;
+import org.apache.flink.util.Preconditions;
+import org.apache.flink.util.ResourceGuard;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import javax.annotation.Nonnull;
+
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Collections;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.NoSuchElementException;
+import java.util.Set;
+import java.util.Spliterators;
+import java.util.TreeSet;
+import java.util.concurrent.ConcurrentHashMap;
+import java.util.concurrent.ThreadLocalRandom;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.stream.Stream;
+import java.util.stream.StreamSupport;
+
+import static org.apache.flink.runtime.state.heap.SkipListUtils.HEAD_NODE;
+import static org.apache.flink.runtime.state.heap.SkipListUtils.NIL_NODE;
+import static 
org.apache.flink.runtime.state.heap.SkipListUtils.NIL_VALUE_POINTER;
+
+/**
+ * Implementation of state map which is based on skip list with copy-on-write 
support. states will
+ * be serialized to bytes and stored in the space allocated with the given 
allocator.
+ *
+ * @param  type of key
+ * @param  type of namespace
+ * @param  type of state
+ */
+public class CopyOnWriteSkipListStateMap extends StateMap 
implements AutoCloseable {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(CopyOnWriteSkipListStateMap.class);
+
+   /**
+* Default max number of logically-removed keys to delete one time.
+*/
+   private static final int DEFAULT_MAX_KEYS_TO_DELETE_ONE_TIME = 3;
+
+   /**
+* Default ratio of the logically-removed keys to trigger deletion when 
snapshot.
+*/
+   private static final float DEFAULT_LOGICAL_REMOVED_KEYS_RATIO = 0.2f;
+
+   /**
+* The serializer used to serialize the key and namespace to bytes 
stored in skip list.
+*/
+   private final SkipListKeySerializer skipListKeySerializer;
+
+   /**
+* The serializer used to serialize the state to bytes stored in skip 
list.
+*/
+   private final SkipListValueSerializer skipListValueSerializer;
+
+   /**
+* Space allocator.
+*/
+   private final Allocator spaceAllocator;
+
+   /**
+* The level index header.
+*/
+   private final LevelIndexHeader levelIndexHeader;
+
+   /**
+* Seed to generate random index level.
+*/
+   private int randomSeed;
+
+   /**
+* The current version of this map. Used for copy-on-write mechanics.
+*/
+   private int stateMapVersion;
+
+   /**
+* The highest version of this map that is still required by any 
unreleased snapshot.
+*/
+   private int highestRequiredSnapshotVersion;
+
+   /**
+* Snapshots no more than this version must have been finished, but 
there may be some
+* snapshots more than this version are still running.
+*/
+   private volatile int highestFinishedSnapshotVersion;
+
+   /**
+* Maintains an ordered set of version ids that are still us

[GitHub] [flink] flinkbot commented on issue #9710: [FLINK-11859][runtime]Improve SpanningRecordSerializer performance by serializing record length to data buffer directly.

2019-09-18 Thread GitBox
flinkbot commented on issue #9710: [FLINK-11859][runtime]Improve 
SpanningRecordSerializer performance by serializing record length to data 
buffer directly.
URL: https://github.com/apache/flink/pull/9710#issuecomment-532961713
 
 
   
   ## CI report:
   
   * 9df06c32a009c53dce3f13c7de12892eb8563213 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9710: [FLINK-11859][runtime]Improve SpanningRecordSerializer performance by serializing record length to data buffer directly.

2019-09-18 Thread GitBox
flinkbot commented on issue #9710: [FLINK-11859][runtime]Improve 
SpanningRecordSerializer performance by serializing record length to data 
buffer directly.
URL: https://github.com/apache/flink/pull/9710#issuecomment-532959252
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 9df06c32a009c53dce3f13c7de12892eb8563213 (Thu Sep 19 
04:22:11 UTC 2019)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-11859) Improve SpanningRecordSerializer performance by serializing record length to serialization buffer directly

2019-09-18 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-11859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao updated FLINK-11859:

Fix Version/s: 1.10.0

> Improve SpanningRecordSerializer performance by serializing record length to 
> serialization buffer directly
> --
>
> Key: FLINK-11859
> URL: https://issues.apache.org/jira/browse/FLINK-11859
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Assignee: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> In the current implementation of SpanningRecordSerializer, the length of a 
> record is serialized to an intermediate length buffer and then copied to the 
> target buffer. Actually, the length filed can be serialized directly to the 
> data buffer (serializationBuffer), which can avoid the copy of length buffer. 
> Though the total bytes copied remain unchanged, it one copy of a small record 
> which incurs high overhead. The flink-benchmarks shows it can improve 
> performance and the test results are as follows.
> Result with the optimization:
> |Benchmark|Mode|Threads|Samples|Score|Score Error (99.9%)|Unit|Param: 
> channelsFlushTimeout|Param: stateBackend|
> |KeyByBenchmarks.arrayKeyBy|thrpt|1|30|2228.049605|77.631804|ops/ms| | |
> |KeyByBenchmarks.tupleKeyBy|thrpt|1|30|3968.361739|193.501755|ops/ms| | |
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|3030.016702|29.272713|ops/ms|
>  |MEMORY|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2754.77678|26.215395|ops/ms|
>  |FS|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|3001.957606|29.288019|ops/ms|
>  |FS_ASYNC|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|123.698984|3.339233|ops/ms|
>  |ROCKS|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|126.252137|1.137735|ops/ms|
>  |ROCKS_INC|
> |SerializationFrameworkMiniBenchmarks.serializerAvro|thrpt|1|30|323.658098|5.855697|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerKryo|thrpt|1|30|183.34423|3.710787|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerPojo|thrpt|1|30|404.380233|5.131744|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerRow|thrpt|1|30|527.193369|10.176726|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerTuple|thrpt|1|30|550.073024|11.724412|ops/ms|
>  | |
> |StreamNetworkBroadcastThroughputBenchmarkExecutor.networkBroadcastThroughput|thrpt|1|30|564.690627|13.766809|ops/ms|
>  | |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|49918.11806|2324.234776|ops/ms|100,100ms|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|10443.63491|315.835962|ops/ms|100,100ms,SSL|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|21387.47608|2779.832704|ops/ms|1000,1ms|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|26585.85453|860.243347|ops/ms|1000,100ms|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|8252.563405|947.129028|ops/ms|1000,100ms,SSL|
>  |
> |SumLongsBenchmark.benchmarkCount|thrpt|1|30|8806.021402|263.995836|ops/ms| | 
> |
> |WindowBenchmarks.globalWindow|thrpt|1|30|4573.620126|112.099391|ops/ms| | |
> |WindowBenchmarks.sessionWindow|thrpt|1|30|585.246412|7.026569|ops/ms| | |
> |WindowBenchmarks.slidingWindow|thrpt|1|30|449.302134|4.123669|ops/ms| | |
> |WindowBenchmarks.tumblingWindow|thrpt|1|30|2979.806858|33.818909|ops/ms| | |
> |StreamNetworkLatencyBenchmarkExecutor.networkLatency1to1|avgt|1|30|12.842865|0.13796|ms/op|
>  | |
> Result without the optimization:
>  
> |Benchmark|Mode|Threads|Samples|Score|Score Error (99.9%)|Unit|Param: 
> channelsFlushTimeout|Param: stateBackend|
> |KeyByBenchmarks.arrayKeyBy|thrpt|1|30|2060.241715|59.898485|ops/ms| | |
> |KeyByBenchmarks.tupleKeyBy|thrpt|1|30|3645.306819|223.821719|ops/ms| | |
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2992.698822|36.978115|ops/ms|
>  |MEMORY|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2756.10949|27.798937|ops/ms|
>  |FS|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2965.969876|44.159793|ops/ms|
>  |FS_ASYNC|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|125.506942|1.245978|ops/ms|
>  |ROCKS|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|127.258737|1.190588|ops/ms|
>  |ROCKS_INC|
> |SerializationFrameworkMiniBenchmarks.serializerAvro|thrpt|1|30|316.497954|8.309241|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerKryo|thrpt|1|30|189.065149|6.302073|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerPojo|thrpt|1|30|391.5

[jira] [Updated] (FLINK-11859) Improve SpanningRecordSerializer performance by serializing record length to serialization buffer directly

2019-09-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-11859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-11859:
---
Labels: pull-request-available  (was: )

> Improve SpanningRecordSerializer performance by serializing record length to 
> serialization buffer directly
> --
>
> Key: FLINK-11859
> URL: https://issues.apache.org/jira/browse/FLINK-11859
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Assignee: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
>
> In the current implementation of SpanningRecordSerializer, the length of a 
> record is serialized to an intermediate length buffer and then copied to the 
> target buffer. Actually, the length filed can be serialized directly to the 
> data buffer (serializationBuffer), which can avoid the copy of length buffer. 
> Though the total bytes copied remain unchanged, it one copy of a small record 
> which incurs high overhead. The flink-benchmarks shows it can improve 
> performance and the test results are as follows.
> Result with the optimization:
> |Benchmark|Mode|Threads|Samples|Score|Score Error (99.9%)|Unit|Param: 
> channelsFlushTimeout|Param: stateBackend|
> |KeyByBenchmarks.arrayKeyBy|thrpt|1|30|2228.049605|77.631804|ops/ms| | |
> |KeyByBenchmarks.tupleKeyBy|thrpt|1|30|3968.361739|193.501755|ops/ms| | |
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|3030.016702|29.272713|ops/ms|
>  |MEMORY|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2754.77678|26.215395|ops/ms|
>  |FS|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|3001.957606|29.288019|ops/ms|
>  |FS_ASYNC|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|123.698984|3.339233|ops/ms|
>  |ROCKS|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|126.252137|1.137735|ops/ms|
>  |ROCKS_INC|
> |SerializationFrameworkMiniBenchmarks.serializerAvro|thrpt|1|30|323.658098|5.855697|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerKryo|thrpt|1|30|183.34423|3.710787|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerPojo|thrpt|1|30|404.380233|5.131744|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerRow|thrpt|1|30|527.193369|10.176726|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerTuple|thrpt|1|30|550.073024|11.724412|ops/ms|
>  | |
> |StreamNetworkBroadcastThroughputBenchmarkExecutor.networkBroadcastThroughput|thrpt|1|30|564.690627|13.766809|ops/ms|
>  | |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|49918.11806|2324.234776|ops/ms|100,100ms|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|10443.63491|315.835962|ops/ms|100,100ms,SSL|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|21387.47608|2779.832704|ops/ms|1000,1ms|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|26585.85453|860.243347|ops/ms|1000,100ms|
>  |
> |StreamNetworkThroughputBenchmarkExecutor.networkThroughput|thrpt|1|30|8252.563405|947.129028|ops/ms|1000,100ms,SSL|
>  |
> |SumLongsBenchmark.benchmarkCount|thrpt|1|30|8806.021402|263.995836|ops/ms| | 
> |
> |WindowBenchmarks.globalWindow|thrpt|1|30|4573.620126|112.099391|ops/ms| | |
> |WindowBenchmarks.sessionWindow|thrpt|1|30|585.246412|7.026569|ops/ms| | |
> |WindowBenchmarks.slidingWindow|thrpt|1|30|449.302134|4.123669|ops/ms| | |
> |WindowBenchmarks.tumblingWindow|thrpt|1|30|2979.806858|33.818909|ops/ms| | |
> |StreamNetworkLatencyBenchmarkExecutor.networkLatency1to1|avgt|1|30|12.842865|0.13796|ms/op|
>  | |
> Result without the optimization:
>  
> |Benchmark|Mode|Threads|Samples|Score|Score Error (99.9%)|Unit|Param: 
> channelsFlushTimeout|Param: stateBackend|
> |KeyByBenchmarks.arrayKeyBy|thrpt|1|30|2060.241715|59.898485|ops/ms| | |
> |KeyByBenchmarks.tupleKeyBy|thrpt|1|30|3645.306819|223.821719|ops/ms| | |
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2992.698822|36.978115|ops/ms|
>  |MEMORY|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2756.10949|27.798937|ops/ms|
>  |FS|
> |MemoryStateBackendBenchmark.stateBackends|thrpt|1|30|2965.969876|44.159793|ops/ms|
>  |FS_ASYNC|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|125.506942|1.245978|ops/ms|
>  |ROCKS|
> |RocksStateBackendBenchmark.stateBackends|thrpt|1|30|127.258737|1.190588|ops/ms|
>  |ROCKS_INC|
> |SerializationFrameworkMiniBenchmarks.serializerAvro|thrpt|1|30|316.497954|8.309241|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerKryo|thrpt|1|30|189.065149|6.302073|ops/ms|
>  | |
> |SerializationFrameworkMiniBenchmarks.serializerPojo|thrpt|1|30|391.51305|7.750728|ops/ms|
>  | |
> |SerializationFrameworkMiniBen

[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128271616)
   * 9abb55b018585e11703afe9f2494ae95fb129c0a : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128277184)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] wsry opened a new pull request #9710: [FLINK-11859][runtime]Improve SpanningRecordSerializer performance by serializing record length to data buffer directly.

2019-09-18 Thread GitBox
wsry opened a new pull request #9710: [FLINK-11859][runtime]Improve 
SpanningRecordSerializer performance by serializing record length to data 
buffer directly.
URL: https://github.com/apache/flink/pull/9710
 
 
   ## What is the purpose of the change
   
   The purpose of this pr is to improve the performance of 
SpanningRecordSerializer. Currently, the data and length filed of a serialized 
record are stored separately in two buffer (the lengthBuffer and the 
serializationBuffer), thus need two times of copy when transferring the 
intermediate data to BufferBuilder. This pr tries to optimize the serialization 
process by removing the lengthBuffer and write the length field to 
serializationBuffer directly, which can avoid the copy of length buffer.
   
   ## Brief change log
   
 - *Remove the length buffer of SpanningRecordSerializer and serialize the 
record length to data buffer directly. More specifically, the initial 4 bytes 
of the data buffer is reserved for length field and after the the serialization 
of record, the reserved space will be filled with record length.*
   
   
   ## Verifying this change
   
- This change is already covered by existing tests, such as 
*SpanningRecordSerializerTest* and *SpanningRecordSerializationTest*.
- The performance gain is proved by the micro-benchmark and the whole 
results can be found in this jira 
[FLINK-11859](https://issues.apache.org/jira/browse/FLINK-11859). 
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (**yes** / no / don't know)
 - The runtime per-record code paths (performance sensitive): (**yes** / no 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / **no** / don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] gaoyunhaii commented on a change in pull request #9679: [hotfix] Fix the comments related to side output.

2019-09-18 Thread GitBox
gaoyunhaii commented on a change in pull request #9679: [hotfix] Fix the 
comments related to side output.
URL: https://github.com/apache/flink/pull/9679#discussion_r325982849
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/OperatorChain.java
 ##
 @@ -549,7 +549,7 @@ public void collect(StreamRecord record) {
@Override
public  void collect(OutputTag outputTag, StreamRecord 
record) {
if (this.outputTag == null || 
!this.outputTag.equals(outputTag)) {
-   // we are only responsible for emitting to the 
side-output specified by our
+   // we are not responsible for emitting to the 
side-output specified by this
 
 Review comment:
   I think if the output tag is null, this output should be responsible for 
emitting to the main output instead of one side output. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] AT-Fieldless opened a new pull request #9350: [FLINK-13485] [chinese-translation] Translate "Table API Example Walkthrough" page into Chinese

2019-09-18 Thread GitBox
AT-Fieldless opened a new pull request #9350: [FLINK-13485] 
[chinese-translation] Translate "Table API Example Walkthrough" page into 
Chinese
URL: https://github.com/apache/flink/pull/9350
 
 
   
   
   ## What is the purpose of the change
   
   Translate "Table API Example Walkthrough" page into Chinese.
   
   
   ## Brief change log
   
   - translate "Table API Example Walkthrough" page into Chinese.
   
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KarmaGYZ commented on a change in pull request #9679: [hotfix] Fix the comments related to side output.

2019-09-18 Thread GitBox
KarmaGYZ commented on a change in pull request #9679: [hotfix] Fix the comments 
related to side output.
URL: https://github.com/apache/flink/pull/9679#discussion_r325982247
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/OperatorChain.java
 ##
 @@ -549,7 +549,7 @@ public void collect(StreamRecord record) {
@Override
public  void collect(OutputTag outputTag, StreamRecord 
record) {
if (this.outputTag == null || 
!this.outputTag.equals(outputTag)) {
-   // we are only responsible for emitting to the 
side-output specified by our
+   // we are not responsible for emitting to the 
side-output specified by this
 
 Review comment:
   I don't think the origin comment is wrong. Same to `RecordWriterOutput.java`


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] xuyang1706 commented on a change in pull request #9184: [FLINK-13339][ml] Add an implementation of pipeline's api

2019-09-18 Thread GitBox
xuyang1706 commented on a change in pull request #9184: [FLINK-13339][ml] Add 
an implementation of pipeline's api
URL: https://github.com/apache/flink/pull/9184#discussion_r325981768
 
 

 ##
 File path: 
flink-ml-parent/flink-ml-lib/src/main/java/org/apache/flink/ml/batchoperator/source/TableSourceBatchOp.java
 ##
 @@ -0,0 +1,44 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
+package org.apache.flink.ml.batchoperator.source;
+
+import org.apache.flink.ml.batchoperator.BatchOperator;
+import org.apache.flink.table.api.Table;
+
+/**
+ * Transform the Table to SourceBatchOp.
+ */
+public final class TableSourceBatchOp extends BatchOperator 
 {
+
+   public TableSourceBatchOp(Table table) {
+   super(null);
+   if (null == table) {
+   throw new RuntimeException();
+   }
+   this.setOutput(table);
+   }
+
+   @Override
+   public TableSourceBatchOp linkFrom(BatchOperator... inputs) {
+   throw new UnsupportedOperationException("Not supported.");
 
 Review comment:
   Thanks, this message is more concise.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] carp84 commented on a change in pull request #9501: [FLINK-12697] [State Backends] Support on-disk state storage for spill-able heap backend

2019-09-18 Thread GitBox
carp84 commented on a change in pull request #9501: [FLINK-12697] [State 
Backends] Support on-disk state storage for spill-able heap backend
URL: https://github.com/apache/flink/pull/9501#discussion_r325980641
 
 

 ##
 File path: 
flink-state-backends/flink-statebackend-heap-spillable/src/main/java/org/apache/flink/runtime/state/heap/CopyOnWriteSkipListStateMap.java
 ##
 @@ -0,0 +1,1527 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.state.heap;
+
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.core.memory.ByteBufferUtils;
+import org.apache.flink.runtime.state.StateEntry;
+import org.apache.flink.runtime.state.StateTransformationFunction;
+import org.apache.flink.runtime.state.heap.space.Allocator;
+import org.apache.flink.runtime.state.heap.space.Chunk;
+import org.apache.flink.runtime.state.heap.space.SpaceUtils;
+import org.apache.flink.runtime.state.internal.InternalKvState;
+import org.apache.flink.util.Preconditions;
+import org.apache.flink.util.ResourceGuard;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import javax.annotation.Nonnull;
+
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Collections;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.NoSuchElementException;
+import java.util.Set;
+import java.util.Spliterators;
+import java.util.TreeSet;
+import java.util.concurrent.ConcurrentHashMap;
+import java.util.concurrent.ThreadLocalRandom;
+import java.util.concurrent.atomic.AtomicBoolean;
+import java.util.stream.Stream;
+import java.util.stream.StreamSupport;
+
+import static org.apache.flink.runtime.state.heap.SkipListUtils.HEAD_NODE;
+import static org.apache.flink.runtime.state.heap.SkipListUtils.NIL_NODE;
+import static 
org.apache.flink.runtime.state.heap.SkipListUtils.NIL_VALUE_POINTER;
+
+/**
+ * Implementation of state map which is based on skip list with copy-on-write 
support. states will
+ * be serialized to bytes and stored in the space allocated with the given 
allocator.
+ *
+ * @param  type of key
+ * @param  type of namespace
+ * @param  type of state
+ */
+public class CopyOnWriteSkipListStateMap extends StateMap 
implements AutoCloseable {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(CopyOnWriteSkipListStateMap.class);
+
+   /**
+* Default max number of logically-removed keys to delete one time.
+*/
+   private static final int DEFAULT_MAX_KEYS_TO_DELETE_ONE_TIME = 3;
+
+   /**
+* Default ratio of the logically-removed keys to trigger deletion when 
snapshot.
+*/
+   private static final float DEFAULT_LOGICAL_REMOVED_KEYS_RATIO = 0.2f;
+
+   /**
+* The serializer used to serialize the key and namespace to bytes 
stored in skip list.
+*/
+   private final SkipListKeySerializer skipListKeySerializer;
+
+   /**
+* The serializer used to serialize the state to bytes stored in skip 
list.
+*/
+   private final SkipListValueSerializer skipListValueSerializer;
+
+   /**
+* Space allocator.
+*/
+   private final Allocator spaceAllocator;
+
+   /**
+* The level index header.
+*/
+   private final LevelIndexHeader levelIndexHeader;
+
+   /**
+* Seed to generate random index level.
+*/
+   private int randomSeed;
+
+   /**
+* The current version of this map. Used for copy-on-write mechanics.
+*/
+   private int stateMapVersion;
+
+   /**
+* The highest version of this map that is still required by any 
unreleased snapshot.
+*/
+   private int highestRequiredSnapshotVersion;
+
+   /**
+* Snapshots no more than this version must have been finished, but 
there may be some
+* snapshots more than this version are still running.
+*/
+   private volatile int highestFinishedSnapshotVersion;
+
+   /**
+* Maintains an ordered set of version ids that are still us

[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128271616)
   * 9abb55b018585e11703afe9f2494ae95fb129c0a : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liupc commented on issue #9683: [FLINK-14037]Use same userCodeClassLoader as jobManager in jobClient to fix input/…

2019-09-18 Thread GitBox
liupc commented on issue #9683: [FLINK-14037]Use same userCodeClassLoader as 
jobManager in jobClient to fix input/…
URL: https://github.com/apache/flink/pull/9683#issuecomment-532955334
 
 
   @TisonKun @aljoscha Thanks for reply, I will update as commented.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zzsmdfj commented on issue #9705: [hotfix][comment]remove useless comment

2019-09-18 Thread GitBox
zzsmdfj commented on issue #9705: [hotfix][comment]remove useless comment
URL: https://github.com/apache/flink/pull/9705#issuecomment-532953571
 
 
   > Thanks for the hotfix, but IMO removed comment is not useless.
   > -1 for this pr.
   
   that comment is just meaning for "session windows UTTests" ? 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on issue #8468: [FLINK-12399][table][table-planner] Fix FilterableTableSource does not change after applyPredicate

2019-09-18 Thread GitBox
KurtYoung commented on issue #8468: [FLINK-12399][table][table-planner] Fix 
FilterableTableSource does not change after applyPredicate
URL: https://github.com/apache/flink/pull/8468#issuecomment-532953494
 
 
   Never mind, it's me didn't explain my thought clear enough.  


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KarmaGYZ commented on issue #9705: [hotfix][comment]remove useless comment

2019-09-18 Thread GitBox
KarmaGYZ commented on issue #9705: [hotfix][comment]remove useless comment
URL: https://github.com/apache/flink/pull/9705#issuecomment-532950384
 
 
   Thanks for the hotfix, but IMO removed comment is not useless.
   -1 for this pr.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14010) Dispatcher & JobManagers don't give up leadership when AM is shut down

2019-09-18 Thread TisonKun (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14010?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933030#comment-16933030
 ] 

TisonKun commented on FLINK-14010:
--

[~trohrmann] Technically I agree that it is a valid solution. Give it another 
look I think we can complete shutdown future exceptionally "ResourceManager got 
closed when DispatcherResourceManagerComponent is running". It infers that the 
application goes into an UNKNOWN state so that the semantic is also correct.

> Dispatcher & JobManagers don't give up leadership when AM is shut down
> --
>
> Key: FLINK-14010
> URL: https://issues.apache.org/jira/browse/FLINK-14010
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN, Runtime / Coordination
>Affects Versions: 1.7.2, 1.8.1, 1.9.0, 1.10.0
>Reporter: TisonKun
>Priority: Critical
>
> In YARN deployment scenario, YARN RM possibly launches a new AM for the job 
> even if the previous AM does not terminated, for example, when AMRM heartbeat 
> timeout. This is a common case that RM will send a shutdown request to the 
> previous AM and expect the AM shutdown properly.
> However, currently in {{YARNResourceManager}}, we handle this request in 
> {{onShutdownRequest}} which simply close the {{YARNResourceManager}} *but not 
> Dispatcher and JobManagers*. Thus, Dispatcher and JobManager launched in new 
> AM cannot be granted leadership properly. Visually,
> on previous AM: Dispatcher leader, JM leaders
> on new AM: ResourceManager leader
> since on client side or in per-job mode, JobManager address and port are 
> configured as the new AM, the whole cluster goes into an unrecoverable 
> inconsistent status: client all queries the dispatcher on new AM who is now 
> the leader. Briefly, Dispatcher and JobManagers on previous AM do not give up 
> their leadership properly.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14118) Reduce the unnecessary flushing when there is no data available for flush

2019-09-18 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933023#comment-16933023
 ] 

Jark Wu commented on FLINK-14118:
-

Is it only affect to 1.9? Should we also fix this in 1.8 and 1.7 series?

cc [~trohrmann]

> Reduce the unnecessary flushing when there is no data available for flush
> -
>
> Key: FLINK-14118
> URL: https://issues.apache.org/jira/browse/FLINK-14118
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0, 1.9.1
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The new flush implementation which works by triggering a netty user event may 
> cause performance regression compared to the old synchronization-based one. 
> More specifically, when there is exactly one BufferConsumer in the buffer 
> queue of subpartition and no new data will be added for a while in the future 
> (may because of just no input or the logic of the operator is to collect some 
> data for processing and will not emit records immediately), that is, there is 
> no data to send, the OutputFlusher will continuously notify data available 
> and wake up the netty thread, though no data will be returned by the 
> pollBuffer method.
> For some of our production jobs, this will incur 20% to 40% CPU overhead 
> compared to the old implementation. We tried to fix the problem by checking 
> if there is new data available when flushing, if there is no new data, the 
> netty thread will not be notified. It works for our jobs and the cpu usage 
> falls to previous level.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14112) Removing zookeeper state should cause the task manager and job managers to restart

2019-09-18 Thread TisonKun (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14112?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933022#comment-16933022
 ] 

TisonKun commented on FLINK-14112:
--

I agree with [~trohrmann]'s comments.

Another question I notice is that for what reason we notify a "null" 
address/session-id? I think the timeout logic can be handled by heartbeats and 
if we enforce the notification always contains valid leader info we can reduce 
noisy & meaningless log also simplify logic in {{LeaderRetrievalListener}}

> Removing zookeeper state should cause the task manager and job managers to 
> restart
> --
>
> Key: FLINK-14112
> URL: https://issues.apache.org/jira/browse/FLINK-14112
> Project: Flink
>  Issue Type: Wish
>  Components: Runtime / Coordination
>Affects Versions: 1.8.1, 1.9.0
>Reporter: Aaron Levin
>Priority: Minor
>
> Suppose you have a flink application running on a cluster with the following 
> configuration:
> {noformat}
> high-availability.zookeeper.path.root: /flink
> {noformat}
> Now suppose you delete all the znodes within {{/flink}}. I experienced the 
> following:
>  * massive amount of logging
>  * application did not restart
>  * task manager did not crash or restart
>  * job manager did not crash or restart
> From this state I had to restart all the task managers and all the job 
> managers in order for the flink application to recover.
> It would be desirable for the Task Managers and Job Managers to crash if the 
> znode is not available (though perhaps you all have thought about this more 
> deeply than I!)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] HuangZhenQiu commented on issue #9689: [FLINK-7151] add a basic function ddl

2019-09-18 Thread GitBox
HuangZhenQiu commented on issue #9689: [FLINK-7151] add a basic function ddl
URL: https://github.com/apache/flink/pull/9689#issuecomment-532946316
 
 
   @suez1224 @fhueske @danny0405 Would you please take a look the PR?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] AT-Fieldless commented on issue #9350: [FLINK-13485] [chinese-translation] Translate "Table API Example Walkthrough" page into Chinese

2019-09-18 Thread GitBox
AT-Fieldless commented on issue #9350: [FLINK-13485] [chinese-translation] 
Translate "Table API Example Walkthrough" page into Chinese
URL: https://github.com/apache/flink/pull/9350#issuecomment-532942135
 
 
   > Hi @AT-Fieldless , do you have time to update this PR?
   
   Yes, I will create a new branch and update this PR.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] AT-Fieldless closed pull request #9350: [FLINK-13485] [chinese-translation] Translate "Table API Example Walkthrough" page into Chinese

2019-09-18 Thread GitBox
AT-Fieldless closed pull request #9350: [FLINK-13485] [chinese-translation] 
Translate "Table API Example Walkthrough" page into Chinese
URL: https://github.com/apache/flink/pull/9350
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] AT-Fieldless commented on issue #9350: [FLINK-13485] [chinese-translation] Translate "Table API Example Walkthrough" page into Chinese

2019-09-18 Thread GitBox
AT-Fieldless commented on issue #9350: [FLINK-13485] [chinese-translation] 
Translate "Table API Example Walkthrough" page into Chinese
URL: https://github.com/apache/flink/pull/9350#issuecomment-532941960
 
 
   
   
   
   > Hi @AT-Fieldless , do you have time to update this PR?
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] AT-Fieldless removed a comment on issue #9350: [FLINK-13485] [chinese-translation] Translate "Table API Example Walkthrough" page into Chinese

2019-09-18 Thread GitBox
AT-Fieldless removed a comment on issue #9350: [FLINK-13485] 
[chinese-translation] Translate "Table API Example Walkthrough" page into 
Chinese
URL: https://github.com/apache/flink/pull/9350#issuecomment-532941960
 
 
   
   
   
   > Hi @AT-Fieldless , do you have time to update this PR?
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (FLINK-14099) SQL supports timestamp in Long

2019-09-18 Thread Zijie Lu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14099?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933009#comment-16933009
 ] 

Zijie Lu edited comment on FLINK-14099 at 9/19/19 2:44 AM:
---

I have tried what you mentioned, but it seems that the timestamp field also 
needs to be standard SQL timestamp and the following exception is throwed. 
[~jark]

 
{quote}Exception in thread "main" org.apache.flink.table.api.TableException: 
findAndCreateTableSource failed.Exception in thread "main" 
org.apache.flink.table.api.TableException: findAndCreateTableSource failed. at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:67)
 at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:54)
 at 
org.apache.flink.table.descriptors.ConnectTableDescriptor.registerTableSource(ConnectTableDescriptor.java:69)
 at sql.job.ms.ClickStats.main(ClickStats.java:73)Caused by: 
org.apache.flink.table.api.ValidationException: Rowtime attribute 'timestamp' 
is not of type SQL_TIMESTAMP. at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceBase.validateRowtimeAttributeDescriptors(KafkaTableSourceBase.java:313)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceBase.(KafkaTableSourceBase.java:118)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSource.(KafkaTableSource.java:66)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactory.createKafkaTableSource(KafkaTableSourceSinkFactory.java:66)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactoryBase.createStreamTableSource(KafkaTableSourceSinkFactoryBase.java:147)
 at 
org.apache.flink.table.factories.StreamTableSourceFactory.createTableSource(StreamTableSourceFactory.java:49)
 at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:65)
 ... 3 more
{quote}


was (Author: alfredlu):
I have tried what you mentioned, but it seems that the timestamp field also 
needs to be standard SQL timestamp and the following exception is throwed.

 
{quote}Exception in thread "main" org.apache.flink.table.api.TableException: 
findAndCreateTableSource failed.Exception in thread "main" 
org.apache.flink.table.api.TableException: findAndCreateTableSource failed. at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:67)
 at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:54)
 at 
org.apache.flink.table.descriptors.ConnectTableDescriptor.registerTableSource(ConnectTableDescriptor.java:69)
 at sql.job.ms.ClickStats.main(ClickStats.java:73)Caused by: 
org.apache.flink.table.api.ValidationException: Rowtime attribute 'timestamp' 
is not of type SQL_TIMESTAMP. at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceBase.validateRowtimeAttributeDescriptors(KafkaTableSourceBase.java:313)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceBase.(KafkaTableSourceBase.java:118)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSource.(KafkaTableSource.java:66)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactory.createKafkaTableSource(KafkaTableSourceSinkFactory.java:66)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactoryBase.createStreamTableSource(KafkaTableSourceSinkFactoryBase.java:147)
 at 
org.apache.flink.table.factories.StreamTableSourceFactory.createTableSource(StreamTableSourceFactory.java:49)
 at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:65)
 ... 3 more
{quote}

> SQL supports timestamp in Long
> --
>
> Key: FLINK-14099
> URL: https://issues.apache.org/jira/browse/FLINK-14099
> Project: Flink
>  Issue Type: New Feature
>Reporter: Zijie Lu
>Priority: Major
>
> The rowtime only supports sql timestamp but not long. Can the rowtime field 
> supports long?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14099) SQL supports timestamp in Long

2019-09-18 Thread Zijie Lu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14099?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16933009#comment-16933009
 ] 

Zijie Lu commented on FLINK-14099:
--

I have tried what you mentioned, but it seems that the timestamp field also 
needs to be standard SQL timestamp and the following exception is throwed.

 
{quote}Exception in thread "main" org.apache.flink.table.api.TableException: 
findAndCreateTableSource failed.Exception in thread "main" 
org.apache.flink.table.api.TableException: findAndCreateTableSource failed. at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:67)
 at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:54)
 at 
org.apache.flink.table.descriptors.ConnectTableDescriptor.registerTableSource(ConnectTableDescriptor.java:69)
 at sql.job.ms.ClickStats.main(ClickStats.java:73)Caused by: 
org.apache.flink.table.api.ValidationException: Rowtime attribute 'timestamp' 
is not of type SQL_TIMESTAMP. at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceBase.validateRowtimeAttributeDescriptors(KafkaTableSourceBase.java:313)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceBase.(KafkaTableSourceBase.java:118)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSource.(KafkaTableSource.java:66)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactory.createKafkaTableSource(KafkaTableSourceSinkFactory.java:66)
 at 
org.apache.flink.streaming.connectors.kafka.KafkaTableSourceSinkFactoryBase.createStreamTableSource(KafkaTableSourceSinkFactoryBase.java:147)
 at 
org.apache.flink.table.factories.StreamTableSourceFactory.createTableSource(StreamTableSourceFactory.java:49)
 at 
org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSource(TableFactoryUtil.java:65)
 ... 3 more
{quote}

> SQL supports timestamp in Long
> --
>
> Key: FLINK-14099
> URL: https://issues.apache.org/jira/browse/FLINK-14099
> Project: Flink
>  Issue Type: New Feature
>Reporter: Zijie Lu
>Priority: Major
>
> The rowtime only supports sql timestamp but not long. Can the rowtime field 
> supports long?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128271616)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14093) Java8 lambdas and exceptions lead to compile error

2019-09-18 Thread zzsmdfj (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14093?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16932996#comment-16932996
 ] 

zzsmdfj commented on FLINK-14093:
-

[~Tison]  yes, you are right. I will start a discussion on dev mailing list. 

> Java8 lambdas and exceptions lead to compile error
> --
>
> Key: FLINK-14093
> URL: https://issues.apache.org/jira/browse/FLINK-14093
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.9.0
>Reporter: zzsmdfj
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> when compile flink master by using Java 1.8.0_77, got errors as follow:
> {code:java}
> // code placeholder
> org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute 
> goal org.apache.maven.plugins:maven-compiler-plugin:3.8.0:
> compile (default-compile) on project flink-table-api-java: Compilation failure
> /home/*/zzsmdfj/sflink/flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/operations/utils/factories/Cal
> culatedTableFactory.java:[90,53] unreported exception X; must be caught or 
> declared to be thrownat 
> org.apache.maven.lifecycle.internal.MojoExecutor.execute 
> (MojoExecutor.java:213)
> at org.apache.maven.lifecycle.internal.MojoExecutor.execute 
> (MojoExecutor.java:154)
> at org.apache.maven.lifecycle.internal.MojoExecutor.execute 
> (MojoExecutor.java:146)
> at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject 
> (LifecycleModuleBuilder.java:117)
> at 
> org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject 
> (LifecycleModuleBuilder.java:81)
> at 
> org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build
>  (SingleThreadedBuilder.java:51)
> at org.apache.maven.lifecycle.internal.LifecycleStarter.execute 
> (LifecycleStarter.java:128)
> at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:309)
> at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:194)
> at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:107)
> at org.apache.maven.cli.MavenCli.execute (MavenCli.java:955)
> at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:290)
> at org.apache.maven.cli.MavenCli.main (MavenCli.java:194)
> at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke 
> (NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke 
> (DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke (Method.java:498)
> at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced 
> (Launcher.java:289)
> at org.codehaus.plexus.classworlds.launcher.Launcher.launch 
> (Launcher.java:229)
> at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode 
> (Launcher.java:415)
> at org.codehaus.plexus.classworlds.launcher.Launcher.main 
> (Launcher.java:356)
> Caused by: org.apache.maven.plugin.compiler.CompilationFailureException: 
> Compilation failure
> {code}
> if using Java 1.8.0_102 to compile, it build success. it maybe a case of bug 
> [JDK-8054569|https://bugs.openjdk.java.net/browse/JDK-8054569], although we 
> can fix this by upgrading the JDK,but i think it is better of compatible with 
> different versions of the JDK8.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14033) Distributed caches are not registered in Yarn Per Job Cluster Mode

2019-09-18 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14033?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16932995#comment-16932995
 ] 

Yang Wang commented on FLINK-14033:
---

[~ZhenqiuHuang]
Yeah, even many users are using -yt to distribute files so that they could be 
accessed by all tasks. We should still support to register by 
`StreamExecutionEnvironment#registerCachedFile`. Just go ahead to fix this. I 
could help to review.

> Distributed caches are not registered in Yarn Per Job Cluster Mode
> --
>
> Key: FLINK-14033
> URL: https://issues.apache.org/jira/browse/FLINK-14033
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.9.0
>Reporter: Zhenqiu Huang
>Assignee: Zhenqiu Huang
>Priority: Minor
>
> CacheFiles in StreamExecutionEnvironment is not used in Job Submission in the 
> Yarn per job cluster mode. Compare to the job submission in session cluster 
> mode that will upload distributed cache files onto http server in application 
> master, we should get the cache files in job graph and register them into 
> blob store in YarnJobClusterEntrypoint.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the StreamingFileSink Builder interface to allow for easier subclassing of StreamingFileSink

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9581: [FLINK-13864][streaming]: Modify the 
StreamingFileSink Builder interface to allow for easier subclassing of 
StreamingFileSink 
URL: https://github.com/apache/flink/pull/9581#issuecomment-526715445
 
 
   
   ## CI report:
   
   * dc9cdc1364b9050c3d2f3cba062310ff68d83b8e : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125277359)
   * 5893967f343a3d2e3b4c92122303e00b01b0f8ff : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125282063)
   * ca8f65756da270da937077a2553a9506066b36bc : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125303705)
   * fc346331b4e7460c75c2e4eb0eefd5f1d1bd7dc9 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/125561291)
   * 3156f2214291b0b18807c50675f7ac6b15e42ba4 : CANCELED 
[Build](https://travis-ci.com/flink-ci/flink/builds/125817723)
   * 66ba043a392e7be15c7aed2f7e2edcaef1472c15 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/125818539)
   * df82b55379806ac8502ef92999a1abc7f9a0056b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/126191145)
   * 9fb1665010c88288eb49c5a9c6d1d2b81e257d71 : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13993) Using FlinkUserCodeClassLoaders to load the user class in the perjob mode

2019-09-18 Thread Guowei Ma (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13993?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guowei Ma updated FLINK-13993:
--
Remaining Estimate: 168h
 Original Estimate: 168h

> Using FlinkUserCodeClassLoaders to load the user class in the perjob mode
> -
>
> Key: FLINK-13993
> URL: https://issues.apache.org/jira/browse/FLINK-13993
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Guowei Ma
>Assignee: Guowei Ma
>Priority: Major
> Fix For: 1.10.0
>
>   Original Estimate: 168h
>  Remaining Estimate: 168h
>
> Currently, Flink has the FlinkUserCodeClassLoader, which is using to load 
> user’s class. However, the user class and the system class are all loaded by 
> the system classloader in the perjob mode. This introduces some conflicts.
> This document[1] gives a proposal that makes the FlinkUserClassLoader load 
> the user class in perjob mode. (disscuss with Till[2])
>  
> [1][https://docs.google.com/document/d/1fH2Cwrrmps5RxxvVuUdeprruvDNabEaIHPyYps28WM8/edit#heading=h.815t5dodlxh7]
> [2] 
> [https://docs.google.com/document/d/1SUhFt1BmsGMLUYVa72SWLbNrrWzunvcjAlEm8iusvq0/edit]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] docete commented on a change in pull request #8706: [FLINK-12814][sql-client] Support a traditional and scrolling view of…

2019-09-18 Thread GitBox
docete commented on a change in pull request #8706: [FLINK-12814][sql-client] 
Support a traditional and scrolling view of…
URL: https://github.com/apache/flink/pull/8706#discussion_r325963191
 
 

 ##
 File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/gateway/local/result/BasicResult.java
 ##
 @@ -50,4 +50,12 @@ public String getWebInterfaceUrl() {
}
return webInterfaceUrl;
}
+
+   /**
+* Result types.
 
 Review comment:
   OK


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] docete commented on a change in pull request #8706: [FLINK-12814][sql-client] Support a traditional and scrolling view of…

2019-09-18 Thread GitBox
docete commented on a change in pull request #8706: [FLINK-12814][sql-client] 
Support a traditional and scrolling view of…
URL: https://github.com/apache/flink/pull/8706#discussion_r325963126
 
 

 ##
 File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/config/entries/ExecutionEntry.java
 ##
 @@ -108,6 +108,10 @@
 
public static final String EXECUTION_CURRENT_DATABASE = 
"current-database";
 
+   private static final String EXECUTION_INTERACTIVE_VIEW_ENABLED = 
"interactive-view.enabled";
+
+   private static final String EXECUTION_MAX_QUERY_TIMEOUT_MS = 
"max-query-timeout-ms";
 
 Review comment:
   It offers an overall protection for the user or downstream projects 
submitting a long-running query(e.g., select * from kafakTable). IMO for DQL 
cases in a streaming scenario, user always expects the result is 
finalizable(He/She can not stare at the screen all of the day). For user, he 
can use CTRL+C signal. For downstream project, a timeout should be better.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-14119) Clean idle state for RetractableTopNFunction

2019-09-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-14119:

Affects Version/s: 1.9.0

> Clean idle state for RetractableTopNFunction
> 
>
> Key: FLINK-14119
> URL: https://issues.apache.org/jira/browse/FLINK-14119
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: Jark Wu
>Assignee: ALEC CHEN
>Priority: Major
>
> We cleaned the idle state for AppendOnlyTopNFunction and 
> UpdatableTopNFunction, but missed this thing for RetractableTopNFunction. We 
> should add it to avoid unlimited state size. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14119) Clean idle state for RetractableTopNFunction

2019-09-18 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14119?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16932988#comment-16932988
 ] 

Jark Wu commented on FLINK-14119:
-

Yes. We need to cleanup the idle state in {{onTimer()}} callback, just like 
UpdatableTopNFunction  did. I assigned this to you.

> Clean idle state for RetractableTopNFunction
> 
>
> Key: FLINK-14119
> URL: https://issues.apache.org/jira/browse/FLINK-14119
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Assignee: ALEC CHEN
>Priority: Major
>
> We cleaned the idle state for AppendOnlyTopNFunction and 
> UpdatableTopNFunction, but missed this thing for RetractableTopNFunction. We 
> should add it to avoid unlimited state size. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-14119) Clean idle state for RetractableTopNFunction

2019-09-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-14119:
---

Assignee: ALEC CHEN

> Clean idle state for RetractableTopNFunction
> 
>
> Key: FLINK-14119
> URL: https://issues.apache.org/jira/browse/FLINK-14119
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Jark Wu
>Assignee: ALEC CHEN
>Priority: Major
>
> We cleaned the idle state for AppendOnlyTopNFunction and 
> UpdatableTopNFunction, but missed this thing for RetractableTopNFunction. We 
> should add it to avoid unlimited state size. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] docete commented on a change in pull request #8706: [FLINK-12814][sql-client] Support a traditional and scrolling view of…

2019-09-18 Thread GitBox
docete commented on a change in pull request #8706: [FLINK-12814][sql-client] 
Support a traditional and scrolling view of…
URL: https://github.com/apache/flink/pull/8706#discussion_r325960076
 
 

 ##
 File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/config/entries/ExecutionEntry.java
 ##
 @@ -108,6 +108,10 @@
 
public static final String EXECUTION_CURRENT_DATABASE = 
"current-database";
 
+   private static final String EXECUTION_INTERACTIVE_VIEW_ENABLED = 
"interactive-view.enabled";
 
 Review comment:
   result-mode describes how we organize the result (change-log or table). and 
this switch describes how we show the result to user(interactive or 
non-interactive). IMO they are two different dimensions.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14038) ExecutionGraph deploy failed due to akka timeout

2019-09-18 Thread liupengcheng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14038?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16932982#comment-16932982
 ] 

liupengcheng commented on FLINK-14038:
--

[~zhuzh] [~trohrmann] This PR not only provide gc logs, but also heapdump, I 
don't think the simple docs in [Getting 
Help|https://flink.apache.org/gettinghelp.html#got-an-error-message]  take 
place of it. In online environment, the time cost is expensive, expecially for 
routine tasks.  If we just provide documents, then user must rerun the 
application,  and also the exceptions may varies, that depends on when OOM or 
long full gc happen. What's more, I think the options is not simple for user to 
get these infos (gc logs, heapdump), but they are basically needed for 
debugging an application.

If there are no drawbacks or regressions, I think we should made it done 
already as an default options for users.

 

> ExecutionGraph deploy failed due to akka timeout
> 
>
> Key: FLINK-14038
> URL: https://issues.apache.org/jira/browse/FLINK-14038
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.9.0
> Environment: Flink on yarn
> Flink 1.9.0
>Reporter: liupengcheng
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> When launching the flink application, the following error was reported, I 
> downloaded the operator logs, but still have no clue. The operator logs 
> provided no useful information and was cancelled directly.
> JobManager logs:
> {code:java}
> java.lang.IllegalStateException: Update task on TaskManager 
> container_e860_1567429198842_571077_01_06 @ zjy-hadoop-prc-st320.bj 
> (dataPort=50990) failed due to:
>   at 
> org.apache.flink.runtime.executiongraph.Execution.lambda$sendUpdatePartitionInfoRpcCall$14(Execution.java:1395)
>   at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>   at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>   at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
>   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:397)
>   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:190)
>   at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
>   at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
>   at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
>   at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
>   at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
>   at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
>   at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
>   at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>   at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>   at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
>   at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
>   at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
>   at akka.actor.ActorCell.invoke(ActorCell.scala:561)
>   at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
>   at akka.dispatch.Mailbox.run(Mailbox.scala:225)
>   at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
>   at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>   at 
> akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>   at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>   at 
> akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> Caused by: java.util.concurrent.CompletionException: 
> akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka.tcp://fl...@zjy-hadoop-prc-st320.bj:62051/user/taskmanager_0#-171547157]]
>  after [1 ms]. Message of type 
> [org.apache.flink.runtime.rpc.messages.RemoteRpcInvocation]. A typical reason 
> for `AskTimeoutException` is that the recipient actor didn't send a reply.
>   at 
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:292)
>   at 
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:308)
>   at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:593)
>   at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>   at 

[GitHub] [flink] docete commented on a change in pull request #8706: [FLINK-12814][sql-client] Support a traditional and scrolling view of…

2019-09-18 Thread GitBox
docete commented on a change in pull request #8706: [FLINK-12814][sql-client] 
Support a traditional and scrolling view of…
URL: https://github.com/apache/flink/pull/8706#discussion_r325958937
 
 

 ##
 File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/CliClient.java
 ##
 @@ -613,6 +624,20 @@ private void printInfo(String message) {
terminal.flush();
}
 
+   /**
+* Output Results in non-interactive way.
+*
+* @return
+*/
+   void printRows(final ResultDescriptor desc, long start, long end) {
 
 Review comment:
   OK


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] docete commented on a change in pull request #8706: [FLINK-12814][sql-client] Support a traditional and scrolling view of…

2019-09-18 Thread GitBox
docete commented on a change in pull request #8706: [FLINK-12814][sql-client] 
Support a traditional and scrolling view of…
URL: https://github.com/apache/flink/pull/8706#discussion_r325959037
 
 

 ##
 File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/CliClient.java
 ##
 @@ -613,6 +624,20 @@ private void printInfo(String message) {
terminal.flush();
}
 
+   /**
+* Output Results in non-interactive way.
+*
+* @return
+*/
+   void printRows(final ResultDescriptor desc, long start, long end) {
+   final String msg = "Time taken: %.03f seconds, Fetched %d 
row(s).";
 
 Review comment:
   OK


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] docete commented on a change in pull request #8706: [FLINK-12814][sql-client] Support a traditional and scrolling view of…

2019-09-18 Thread GitBox
docete commented on a change in pull request #8706: [FLINK-12814][sql-client] 
Support a traditional and scrolling view of…
URL: https://github.com/apache/flink/pull/8706#discussion_r325958597
 
 

 ##
 File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/CliClient.java
 ##
 @@ -464,28 +465,38 @@ private void callExplain(SqlCommandCall cmdCall) {
}
 
private void callSelect(SqlCommandCall cmdCall) {
+   final long start = System.currentTimeMillis();
final ResultDescriptor resultDesc;
try {
resultDesc = executor.executeQuery(context, 
cmdCall.operands[0]);
} catch (SqlExecutionException e) {
printExecutionException(e);
return;
}
-   final CliResultView view;
-   if (resultDesc.isMaterialized()) {
-   view = new CliTableResultView(this, resultDesc);
-   } else {
-   view = new CliChangelogResultView(this, resultDesc);
-   }
+   final long end = System.currentTimeMillis();
 
 Review comment:
   User can wait a timeout(setting by max-query-timeout-ms) or just type CTRL + 
C to terminate the submitted query.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-14122) Extend State Processor API to read ListCheckpointed operator state

2019-09-18 Thread Seth Wiesman (Jira)
Seth Wiesman created FLINK-14122:


 Summary: Extend State Processor API to read ListCheckpointed 
operator state
 Key: FLINK-14122
 URL: https://issues.apache.org/jira/browse/FLINK-14122
 Project: Flink
  Issue Type: Sub-task
  Components: API / DataStream
Affects Versions: 1.10.0
Reporter: Seth Wiesman


The state processor api cannot  read operator state using the ListCheckpointed 
interface because it requires access the JavaSerializer which is package 
private. Instead of making that class public, we should offer a 
readListCheckpointed Method to easily read this state.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] TisonKun commented on issue #9684: [FLINK-13992][coordination] Refactor Optional parameter in InputGateWithMetrics#updateMetrics

2019-09-18 Thread GitBox
TisonKun commented on issue #9684: [FLINK-13992][coordination] Refactor 
Optional parameter in InputGateWithMetrics#updateMetrics
URL: https://github.com/apache/flink/pull/9684#issuecomment-532922137
 
 
   ping @azagrebin as a reminder :-)


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14109) Improve javadocs and tests for high-availability backend

2019-09-18 Thread TisonKun (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16932959#comment-16932959
 ] 

TisonKun commented on FLINK-14109:
--

[~bremac] aha, nice to hear :-)

> Improve javadocs and tests for high-availability backend
> 
>
> Key: FLINK-14109
> URL: https://issues.apache.org/jira/browse/FLINK-14109
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Reporter: Brendan MacDonell
>Assignee: Brendan MacDonell
>Priority: Minor
>
> My team at Sight Machine just finished building a custom HA backend for our 
> infrastructure. The process was mostly painless, but we ran into a few pain 
> points on the way:
>  * {{CompletedCheckpointStore#getLatestCheckpoint}} is not marked as 
> {{@Nullable}}, so there was some confusion about whether the Javadocs or 
> (lack of) annotation is correct. The interface would be clearer if the 
> annotation was present.
>  * The javadocs for {{CompletedCheckpointStore#recover}} disagree with the 
> documentation for {{ZooKeeperCompletedCheckpointStore#recover}}. It's not 
> immediately clear to someone working on the code that the ZK javadoc is 
> outdated and the interface documentation is correct.
>  * -The base {{CompletedCheckpointStore}} tests only work with 
> high-availability backends that keep a list of checkpoints in memory. If the 
> backend persists and retrieves data from another source the tests will fail. 
> It's fairly simple to change the tests to lift this requirement though. See 
> [this gist|https://gist.github.com/bremac/1b3365bc0257dfbd33bcd0b7a7627c00] 
> for an example.-
> We've got patches for the points above that we'd be happy to contribute. :)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] tweise merged pull request #9708: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
tweise merged pull request #9708: [FLINK-14107][kinesis] Erroneous queue 
selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9708
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] tweise commented on issue #9708: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
tweise commented on issue #9708: [FLINK-14107][kinesis] Erroneous queue 
selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9708#issuecomment-532910734
 
 
   unrelated test failure


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] tweise merged pull request #9709: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
tweise merged pull request #9709: [FLINK-14107][kinesis] Erroneous queue 
selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9709
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] tweise commented on issue #9709: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
tweise commented on issue #9709: [FLINK-14107][kinesis] Erroneous queue 
selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9709#issuecomment-532910374
 
 
   unrelated test failure


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9708: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9708: [FLINK-14107][kinesis] Erroneous 
queue selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9708#issuecomment-532891597
 
 
   
   ## CI report:
   
   * adf5d969d37c127653d77821bfa5fe48cf212a6d : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128256147)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9709: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9709: [FLINK-14107][kinesis] Erroneous 
queue selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9709#issuecomment-532891619
 
 
   
   ## CI report:
   
   * 02a0e0ea41b1dc31f233ca4d25434b5370451d1a : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/128256191)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9709: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9709: [FLINK-14107][kinesis] Erroneous 
queue selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9709#issuecomment-532891619
 
 
   
   ## CI report:
   
   * 02a0e0ea41b1dc31f233ca4d25434b5370451d1a : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128256191)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9708: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
flinkbot edited a comment on issue #9708: [FLINK-14107][kinesis] Erroneous 
queue selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9708#issuecomment-532891597
 
 
   
   ## CI report:
   
   * adf5d969d37c127653d77821bfa5fe48cf212a6d : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/128256147)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9708: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
flinkbot commented on issue #9708: [FLINK-14107][kinesis] Erroneous queue 
selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9708#issuecomment-532891597
 
 
   
   ## CI report:
   
   * adf5d969d37c127653d77821bfa5fe48cf212a6d : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9709: [FLINK-14107][kinesis] Erroneous queue selection in record emitter may lead to deadlock

2019-09-18 Thread GitBox
flinkbot commented on issue #9709: [FLINK-14107][kinesis] Erroneous queue 
selection in record emitter may lead to deadlock
URL: https://github.com/apache/flink/pull/9709#issuecomment-532891619
 
 
   
   ## CI report:
   
   * 02a0e0ea41b1dc31f233ca4d25434b5370451d1a : UNKNOWN
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Resolved] (FLINK-14107) Kinesis consumer record emitter deadlock under event time alignment

2019-09-18 Thread Thomas Weise (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Weise resolved FLINK-14107.
--
Fix Version/s: 1.10.0
   Resolution: Fixed

> Kinesis consumer record emitter deadlock under event time alignment
> ---
>
> Key: FLINK-14107
> URL: https://issues.apache.org/jira/browse/FLINK-14107
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kinesis
>Affects Versions: 1.8.2, 1.9.0
>Reporter: Thomas Weise
>Assignee: Thomas Weise
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> When the emitter reaches the max timestamp for the current queue, it stops 
> emitting and waits for the max timestamp to advance. Since it simultaneously 
> selects the next queue as the new "minimum" queue, it may deadlock if the 
> previous min queue represents the new global lower bound after the max 
> timestamp advanced. This occurs very infrequently and we were finally able to 
> reproduce.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   3   4   >