[jira] [Created] (FLINK-15323) SQL Client end-to-end test (Old planner) failed on travis

2019-12-18 Thread Yu Li (Jira)
Yu Li created FLINK-15323:
-

 Summary: SQL Client end-to-end test (Old planner) failed on travis
 Key: FLINK-15323
 URL: https://issues.apache.org/jira/browse/FLINK-15323
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Legacy Planner
Affects Versions: 1.10.0
Reporter: Yu Li
 Fix For: 1.10.0


As titled, the 'SQL Client end-to-end test (Old planner)' case failed with 
below error:
{code}
org.apache.flink.table.client.gateway.SqlExecutionException: Invalid SQL update 
statement.
at 
org.apache.flink.table.client.gateway.local.LocalExecutor.applyUpdate(LocalExecutor.java:697)
at 
org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdateInternal(LocalExecutor.java:576)
at 
org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdate(LocalExecutor.java:527)
at 
org.apache.flink.table.client.cli.CliClient.callInsertInto(CliClient.java:535)
at 
org.apache.flink.table.client.cli.CliClient.lambda$submitUpdate$0(CliClient.java:231)
at java.util.Optional.map(Optional.java:215)
at 
org.apache.flink.table.client.cli.CliClient.submitUpdate(CliClient.java:228)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:129)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:178)
Caused by: java.lang.RuntimeException: Error while applying rule 
PushProjectIntoTableSourceScanRule, args 
[rel#88:FlinkLogicalCalc.LOGICAL(input=RelSubset#87,expr#0..2={inputs},expr#3=IS
 NOT NULL($t1),user=$t1,rowtime=$t0,$condition=$t3), 
Scan(table:[default_catalog, default_database, JsonSourceTable], 
fields:(rowtime, user, event), source:KafkaTableSource(rowtime, user, event))]
at 
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:235)
at 
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:631)
at 
org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:327)
at 
org.apache.flink.table.plan.Optimizer.runVolcanoPlanner(Optimizer.scala:280)
at 
org.apache.flink.table.plan.Optimizer.optimizeLogicalPlan(Optimizer.scala:199)
at 
org.apache.flink.table.plan.StreamOptimizer.optimize(StreamOptimizer.scala:66)
at 
org.apache.flink.table.planner.StreamPlanner.writeToUpsertSink(StreamPlanner.scala:353)
at 
org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$writeToSink(StreamPlanner.scala:281)
...
Caused by: org.apache.flink.table.api.ValidationException: Rowtime field 
'rowtime' has invalid type LocalDateTime. Rowtime attributes must be of type 
Timestamp.
at 
org.apache.flink.table.sources.TableSourceUtil$$anonfun$3.apply(TableSourceUtil.scala:114)
at 
org.apache.flink.table.sources.TableSourceUtil$$anonfun$3.apply(TableSourceUtil.scala:92)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at 
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at 
scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
at scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186)
at 
org.apache.flink.table.sources.TableSourceUtil$.computeIndexMapping(TableSourceUtil.scala:92)
{code}

https://api.travis-ci.org/v3/job/626790280/log.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15320) JobManager crashes in the standalone model when cancelling job which subtask' status is scheduled

2019-12-18 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15320?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999815#comment-16999815
 ] 

Zhu Zhu commented on FLINK-15320:
-

To summarize, this issue happens when an external job cancel request comes when 
the job is still allocating slots.
The root cause is that, the version of the canceled vertices are not 
incremented in the case of external job cancel request, and the pending slot 
requests are also not canceled in this case, so that the returned slot can be 
used to fulfill an outdated deployment, which finally triggers the fatal error.
To fix it, I think we should always increment the version of a vertex before 
canceling/failing it. Besides {{JobMaster#cancel}}, there are several other 
cases including {{JobMaster#suspend}} and {{ExecutionGraph#failJob}}.

cc [~gjy]

> JobManager crashes in the standalone model when cancelling job which subtask' 
> status is scheduled
> -
>
> Key: FLINK-15320
> URL: https://issues.apache.org/jira/browse/FLINK-15320
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.10.0
>Reporter: lining
>Assignee: Zhu Zhu
>Priority: Blocker
> Fix For: 1.10.0
>
>
> Use start-cluster.sh to start a standalone cluster, and then submit a job 
> from the streaming's example which name is TopSpeedWindowing, parallelism is 
> 20. Wait for one minute, cancel the job, jobmanager will crash. The exception 
> stack is:
> 2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
> Could not assign resource 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
> current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
> TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print 
> to Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
>  at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
>  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) 
> at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
>  at 
> 

[jira] [Updated] (FLINK-15308) Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1

2019-12-18 Thread zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15308?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhijiang updated FLINK-15308:
-
Fix Version/s: 1.10.0

> Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1
> --
>
> Key: FLINK-15308
> URL: https://issues.apache.org/jira/browse/FLINK-15308
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.0
> Environment: $ git log
> commit 4b54da2c67692b1c9d43e1184c00899b0151b3ae
> Author: bowen.li 
> Date: Tue Dec 17 17:37:03 2019 -0800
>Reporter: Feng Jiajie
>Assignee: Yingjie Cao
>Priority: Blocker
> Fix For: 1.10.0
>
> Attachments: image-2019-12-19-10-55-30-644.png
>
>
> Job worked well with default flink-conf.yaml with 
> pipelined-shuffle.compression:
> {code:java}
> taskmanager.numberOfTaskSlots: 1
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> But when I set taskmanager.numberOfTaskSlots to 4 or 6:
> {code:java}
> taskmanager.numberOfTaskSlots: 6
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> job failed:
> {code:java}
> $ bin/flink run -m yarn-cluster -p 16 -yjm 1024m -ytm 12288m 
> ~/flink-example-1.0-SNAPSHOT.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/lib/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/data/sa_cluster/cloudera/parcels/CDH-5.14.4-1.cdh5.14.4.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,907 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - No path for the flink jar passed. Using the location of class 
> org.apache.flink.yarn.YarnClusterDescriptor to locate the jar
> 2019-12-18 15:04:41,084 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Cluster specification: 
> ClusterSpecification{masterMemoryMB=1024, taskManagerMemoryMB=12288, 
> numberTaskManagers=1, slotsPerTaskManager=6}
> 2019-12-18 15:04:42,344 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Submitting application master application_1576573857638_0026
> 2019-12-18 15:04:42,370 INFO  
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted 
> application application_1576573857638_0026
> 2019-12-18 15:04:42,371 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Waiting for the cluster to be allocated
> 2019-12-18 15:04:42,372 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Deploying cluster, current state ACCEPTED
> 2019-12-18 15:04:45,388 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - YARN application has been deployed successfully.
> 2019-12-18 15:04:45,390 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Found Web Interface debugboxcreate431x3.sa:36162 of 
> application 'application_1576573857638_0026'.
> Job has been submitted with JobID 9140c70769f4271cc22ea8becaa26272
> 
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: org.apache.flink.client.program.ProgramInvocationException: 
> Job failed (JobID: 9140c70769f4271cc22ea8becaa26272)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>   at 
> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>   at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:664)
>   at 

[GitHub] [flink] lirui-apache opened a new pull request #10625: [FLINK-15259][hive] HiveInspector.toInspectors() should convert Flink…

2019-12-18 Thread GitBox
lirui-apache opened a new pull request #10625: [FLINK-15259][hive] 
HiveInspector.toInspectors() should convert Flink…
URL: https://github.com/apache/flink/pull/10625
 
 
   … constant to Hive constant
   
   
   
   ## What is the purpose of the change
   
   Convert constant Flink object to Hive object before creating Hive's constant 
object inspectors.
   
   
   ## Brief change log
   
 - Convert the constant value to Hive object in 
`HiveInspectors::toInspectors`
 - Implement our own java constant objector inspectors for char, varchar 
and decimal
 - Add test case
   
   
   ## Verifying this change
   
   New test case
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? NA
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15259) HiveInspector.toInspectors() should convert Flink constant to Hive constant

2019-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-15259:
---
Labels: pull-request-available  (was: )

> HiveInspector.toInspectors() should convert Flink constant to Hive constant 
> 
>
> Key: FLINK-15259
> URL: https://issues.apache.org/jira/browse/FLINK-15259
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Bowen Li
>Assignee: Rui Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.9.2, 1.10.0, 1.11.0
>
>
> repro test: 
> {code:java}
> public class HiveModuleITCase {
>   @Test
>   public void test() {
>   TableEnvironment tEnv = 
> HiveTestUtils.createTableEnvWithBlinkPlannerBatchMode();
>   tEnv.unloadModule("core");
>   tEnv.loadModule("hive", new HiveModule("2.3.4"));
>   tEnv.sqlQuery("select concat('an', 'bn')");
>   }
> }
> {code}
> seems that currently HiveInspector.toInspectors() didn't convert Flink 
> constant to Hive constant before calling 
> hiveShim.getObjectInspectorForConstant
> I don't think it's a blocker



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-15311) Lz4BlockCompressionFactory should use native compressor instead of java unsafe

2019-12-18 Thread zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhijiang reassigned FLINK-15311:


Assignee: Yingjie Cao

> Lz4BlockCompressionFactory should use native compressor instead of java unsafe
> --
>
> Key: FLINK-15311
> URL: https://issues.apache.org/jira/browse/FLINK-15311
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Reporter: Jingsong Lee
>Assignee: Yingjie Cao
>Priority: Critical
> Fix For: 1.10.0
>
>
> According to:
> [https://lz4.github.io/lz4-java/1.7.0/lz4-compression-benchmark/]
> Java java unsafe compressor has lower performance than native lz4 compressor.
> After FLINK-14845 , we use lz4 compression for shuffler.
> In testing, I found shuffle using java unsafe compressor.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15311) Lz4BlockCompressionFactory should use native compressor instead of java unsafe

2019-12-18 Thread zhijiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999797#comment-16999797
 ] 

zhijiang commented on FLINK-15311:
--

I guess it belongs to performance improvement, not a bug, because it does not 
affect the compression function and stable issue.

Should it be a blocker for the release?

> Lz4BlockCompressionFactory should use native compressor instead of java unsafe
> --
>
> Key: FLINK-15311
> URL: https://issues.apache.org/jira/browse/FLINK-15311
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Reporter: Jingsong Lee
>Priority: Critical
> Fix For: 1.10.0
>
>
> According to:
> [https://lz4.github.io/lz4-java/1.7.0/lz4-compression-benchmark/]
> Java java unsafe compressor has lower performance than native lz4 compressor.
> After FLINK-14845 , we use lz4 compression for shuffler.
> In testing, I found shuffle using java unsafe compressor.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15199) Benchmarks are not compiling

2019-12-18 Thread Yu Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999791#comment-16999791
 ] 

Yu Li commented on FLINK-15199:
---

Great to know, thanks for the update! [~pnowojski]

> Benchmarks are not compiling
> 
>
> Key: FLINK-15199
> URL: https://issues.apache.org/jira/browse/FLINK-15199
> Project: Flink
>  Issue Type: Bug
>  Components: Benchmarks
>Affects Versions: 1.10.0
>Reporter: Piotr Nowojski
>Assignee: Piotr Nowojski
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Recent changes in FLINK-14926 caused:
> {noformat}
> [ERROR] COMPILATION ERROR : 
> [INFO] -
> [ERROR] 
> /home/jenkins/workspace/flink-master-benchmarks/flink-benchmarks/src/main/java/org/apache/flink/state/benchmark/BackendUtils.java:[61,57]
>  cannot infer type arguments for 
> org.apache.flink.contrib.streaming.state.RocksDBKeyedStateBackendBuilder<>
> [INFO] 1 error
> [INFO] -
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> [INFO] Total time: 2.635 s
> [INFO] Finished at: 2019-12-11T14:58:37+01:00
> [INFO] Final Memory: 31M/751M
> [INFO] 
> 
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) 
> on project flink-hackathon-benchmarks: Compilation failure
> [ERROR] 
> /home/jenkins/workspace/flink-master-benchmarks/flink-benchmarks/src/main/java/org/apache/flink/state/benchmark/BackendUtils.java:[61,57]
>  cannot infer type arguments for 
> org.apache.flink.contrib.streaming.state.RocksDBKeyedStateBackendBuilder<>
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15322) Parquet test fails with Hive versions prior to 1.2.0

2019-12-18 Thread Rui Li (Jira)
Rui Li created FLINK-15322:
--

 Summary: Parquet test fails with Hive versions prior to 1.2.0
 Key: FLINK-15322
 URL: https://issues.apache.org/jira/browse/FLINK-15322
 Project: Flink
  Issue Type: Test
  Components: Connectors / Hive
Reporter: Rui Li
 Fix For: 1.10.0


Some data types are not supported by parquet tables in older Hive versions. 
Related to: https://issues.apache.org/jira/browse/HIVE-6384



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15320) JobManager crashes in the standalone model when cancelling job which subtask' status is scheduled

2019-12-18 Thread Yu Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yu Li updated FLINK-15320:
--
Fix Version/s: 1.10.0

> JobManager crashes in the standalone model when cancelling job which subtask' 
> status is scheduled
> -
>
> Key: FLINK-15320
> URL: https://issues.apache.org/jira/browse/FLINK-15320
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.10.0
>Reporter: lining
>Assignee: Zhu Zhu
>Priority: Blocker
> Fix For: 1.10.0
>
>
> Use start-cluster.sh to start a standalone cluster, and then submit a job 
> from the streaming's example which name is TopSpeedWindowing, parallelism is 
> 20. Wait for one minute, cancel the job, jobmanager will crash. The exception 
> stack is:
> 2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
> Could not assign resource 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
> current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
> TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print 
> to Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
>  at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
>  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) 
> at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.releaseSharedSlot(SchedulerImpl.java:552)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.cancelSlotRequest(SchedulerImpl.java:184)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.returnLogicalSlot(SchedulerImpl.java:195)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.lambda$returnSlotToOwner$0(SingleLogicalSlot.java:181)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenCompleteStage(CompletableFuture.java:778)
>  at 
> 

[jira] [Assigned] (FLINK-15320) JobManager crashes in the standalone model when cancelling job which subtask' status is scheduled

2019-12-18 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu reassigned FLINK-15320:
---

Assignee: Zhu Zhu

> JobManager crashes in the standalone model when cancelling job which subtask' 
> status is scheduled
> -
>
> Key: FLINK-15320
> URL: https://issues.apache.org/jira/browse/FLINK-15320
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.10.0
>Reporter: lining
>Assignee: Zhu Zhu
>Priority: Blocker
>
> Use start-cluster.sh to start a standalone cluster, and then submit a job 
> from the streaming's example which name is TopSpeedWindowing, parallelism is 
> 20. Wait for one minute, cancel the job, jobmanager will crash. The exception 
> stack is:
> 2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
> Could not assign resource 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
> current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
> TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print 
> to Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
>  at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
>  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) 
> at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.releaseSharedSlot(SchedulerImpl.java:552)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.cancelSlotRequest(SchedulerImpl.java:184)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.returnLogicalSlot(SchedulerImpl.java:195)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.lambda$returnSlotToOwner$0(SingleLogicalSlot.java:181)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenCompleteStage(CompletableFuture.java:778)
>  at 
> java.util.concurrent.CompletableFuture.whenComplete(CompletableFuture.java:2140)
>  at 
> 

[jira] [Commented] (FLINK-15258) HiveModuleFactory should be able to distinguish hive-version as property

2019-12-18 Thread Bowen Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999762#comment-16999762
 ] 

Bowen Li commented on FLINK-15258:
--

[~trohrmann] I thought the original title was pretty self explanatary, it 
probably can be rephrased as "HiveModuleFactory should be able to distinguish 
hive-version as property". Thanks for the reminder

> HiveModuleFactory should be able to distinguish hive-version as property
> 
>
> Key: FLINK-15258
> URL: https://issues.apache.org/jira/browse/FLINK-15258
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.10.0
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Blocker
> Fix For: 1.10.0, 1.11.0
>
>
> HiveModuleFactory should be able to distinguish hive-version as supported 
> property. it currently can't



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15258) HiveModuleFactory should be able to distinguish hive-version as property

2019-12-18 Thread Bowen Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-15258:
-
Description: HiveModuleFactory should be able to distinguish hive-version 
as supported property. it currently can't  (was: HiveModuleFactory should be 
able to distinguish hive-version as supported property)

> HiveModuleFactory should be able to distinguish hive-version as property
> 
>
> Key: FLINK-15258
> URL: https://issues.apache.org/jira/browse/FLINK-15258
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.10.0
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Blocker
> Fix For: 1.10.0, 1.11.0
>
>
> HiveModuleFactory should be able to distinguish hive-version as supported 
> property. it currently can't



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15258) HiveModuleFactory should take hive-version as property

2019-12-18 Thread Bowen Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-15258:
-
Description: HiveModuleFactory should be able to distinguish hive-version 
as supported property  (was: HiveModuleFactory should have hive-version as 
supported property)

> HiveModuleFactory should take hive-version as property
> --
>
> Key: FLINK-15258
> URL: https://issues.apache.org/jira/browse/FLINK-15258
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.10.0
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Blocker
> Fix For: 1.10.0, 1.11.0
>
>
> HiveModuleFactory should be able to distinguish hive-version as supported 
> property



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15258) HiveModuleFactory should be able to distinguish hive-version as property

2019-12-18 Thread Bowen Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li updated FLINK-15258:
-
Summary: HiveModuleFactory should be able to distinguish hive-version as 
property  (was: HiveModuleFactory should take hive-version as property)

> HiveModuleFactory should be able to distinguish hive-version as property
> 
>
> Key: FLINK-15258
> URL: https://issues.apache.org/jira/browse/FLINK-15258
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.10.0
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Blocker
> Fix For: 1.10.0, 1.11.0
>
>
> HiveModuleFactory should be able to distinguish hive-version as supported 
> property



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread hehuiyuan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999722#comment-16999722
 ] 

hehuiyuan edited comment on FLINK-15321 at 12/19/19 3:58 AM:
-

You can try 

{{CONCAT_WS()}}


was (Author: hehuiyuan):
You can try 

{{CONCAT_WS(string1, string2, string3,...)}}

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread hehuiyuan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999722#comment-16999722
 ] 

hehuiyuan commented on FLINK-15321:
---

You can try 

{{CONCAT_WS(string1, string2, string3,...)}}

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread xiaojin.wy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999718#comment-16999718
 ] 

xiaojin.wy edited comment on FLINK-15321 at 12/19/19 3:50 AM:
--

Ok, I will close it if it's expected.


was (Author: xiaojin.wy):
ok, I will close it if it's expected.

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread xiaojin.wy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999718#comment-16999718
 ] 

xiaojin.wy commented on FLINK-15321:


ok, I will close it if it's expected.

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-15321.
---
Fix Version/s: (was: 1.10.0)
   Resolution: Not A Bug

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999716#comment-16999716
 ] 

Jark Wu commented on FLINK-15321:
-

This is as expected. {{concat}} will return null if any argument is null. see 
the documentation [1].

[1]: 
https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/systemFunctions.html

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
> Fix For: 1.10.0
>
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-15321:

Fix Version/s: 1.10.0

> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
> Fix For: 1.10.0
>
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread xiaojin.wy (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15321?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

xiaojin.wy updated FLINK-15321:
---
Description: 
The sql is :
SELECT concat('a', cast(null as varchar), 'c');

The result of the sql is NULL after you execute it in a sqlClient 
environment;But actually the result should be 'ac'.

The config is:
execution:
  planner: blink
  type: batch

  was:
The sql is :
SELECT concat('a', cast(null as varchar), 'c');

The result of the sql is NULL after you execute it in a sqlClient 
environment;But actually the result should be 'ac'.


> The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;
> ---
>
> Key: FLINK-15321
> URL: https://issues.apache.org/jira/browse/FLINK-15321
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
>
> The sql is :
> SELECT concat('a', cast(null as varchar), 'c');
> The result of the sql is NULL after you execute it in a sqlClient 
> environment;But actually the result should be 'ac'.
> The config is:
> execution:
>   planner: blink
>   type: batch



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15321) The result of sql(SELECT concat('a', cast(null as varchar), 'c');) is NULL;

2019-12-18 Thread xiaojin.wy (Jira)
xiaojin.wy created FLINK-15321:
--

 Summary: The result of sql(SELECT concat('a', cast(null as 
varchar), 'c');) is NULL;
 Key: FLINK-15321
 URL: https://issues.apache.org/jira/browse/FLINK-15321
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Client
Affects Versions: 1.10.0
Reporter: xiaojin.wy


The sql is :
SELECT concat('a', cast(null as varchar), 'c');

The result of the sql is NULL after you execute it in a sqlClient 
environment;But actually the result should be 'ac'.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15310) A timestamp result get by a select sql and a csvsink sql is different

2019-12-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15310?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-15310:

Fix Version/s: 1.10.0

> A timestamp result get by a select sql and a csvsink sql is different
> -
>
> Key: FLINK-15310
> URL: https://issues.apache.org/jira/browse/FLINK-15310
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
> Environment: execution:
>   planner: blink
>   type: batch
>Reporter: xiaojin.wy
>Priority: Major
> Fix For: 1.10.0
>
> Attachments: image-2019-12-18-16-33-24-510.png
>
>
> *The sql is:*
> CREATE TABLE `orders` (
>   rowtime TIMESTAMP,
>   id  INT,
>   product VARCHAR,
>   units INT
> ) WITH (
>   'format.field-delimiter'='|',
>   'connector.type'='filesystem',
>   'format.derive-schema'='true',
>   
> 'connector.path'='/defender_test_data/daily_regression_blink_sql_1.10/test_agg/sources/orders.csv',
>   'format.type'='csv'
> );
> select floor(rowtime to hour) as rowtime, count(*) as c from orders group by 
> floor(rowtime to hour)
> The result got in a sqlClient environment which use the sql above is like 
> this:
>  !image-2019-12-18-16-33-24-510.png! 
> But the same sql write directly to a cvs batch sink will get a result like 
> this:
> 1972-03-06 08:44:36.736|4
> 1972-03-06 08:44:36.736|1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15310) A timestamp result get by a select sql and a csvsink sql is different

2019-12-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15310?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-15310:

Priority: Critical  (was: Major)

> A timestamp result get by a select sql and a csvsink sql is different
> -
>
> Key: FLINK-15310
> URL: https://issues.apache.org/jira/browse/FLINK-15310
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
> Environment: execution:
>   planner: blink
>   type: batch
>Reporter: xiaojin.wy
>Priority: Critical
> Fix For: 1.10.0
>
> Attachments: image-2019-12-18-16-33-24-510.png
>
>
> *The sql is:*
> CREATE TABLE `orders` (
>   rowtime TIMESTAMP,
>   id  INT,
>   product VARCHAR,
>   units INT
> ) WITH (
>   'format.field-delimiter'='|',
>   'connector.type'='filesystem',
>   'format.derive-schema'='true',
>   
> 'connector.path'='/defender_test_data/daily_regression_blink_sql_1.10/test_agg/sources/orders.csv',
>   'format.type'='csv'
> );
> select floor(rowtime to hour) as rowtime, count(*) as c from orders group by 
> floor(rowtime to hour)
> The result got in a sqlClient environment which use the sql above is like 
> this:
>  !image-2019-12-18-16-33-24-510.png! 
> But the same sql write directly to a cvs batch sink will get a result like 
> this:
> 1972-03-06 08:44:36.736|4
> 1972-03-06 08:44:36.736|1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15308) Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1

2019-12-18 Thread Yingjie Cao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999708#comment-16999708
 ] 

Yingjie Cao commented on FLINK-15308:
-

[~fengjiajie] I also reproduced it.

> Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1
> --
>
> Key: FLINK-15308
> URL: https://issues.apache.org/jira/browse/FLINK-15308
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.0
> Environment: $ git log
> commit 4b54da2c67692b1c9d43e1184c00899b0151b3ae
> Author: bowen.li 
> Date: Tue Dec 17 17:37:03 2019 -0800
>Reporter: Feng Jiajie
>Assignee: Yingjie Cao
>Priority: Blocker
> Attachments: image-2019-12-19-10-55-30-644.png
>
>
> Job worked well with default flink-conf.yaml with 
> pipelined-shuffle.compression:
> {code:java}
> taskmanager.numberOfTaskSlots: 1
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> But when I set taskmanager.numberOfTaskSlots to 4 or 6:
> {code:java}
> taskmanager.numberOfTaskSlots: 6
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> job failed:
> {code:java}
> $ bin/flink run -m yarn-cluster -p 16 -yjm 1024m -ytm 12288m 
> ~/flink-example-1.0-SNAPSHOT.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/lib/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/data/sa_cluster/cloudera/parcels/CDH-5.14.4-1.cdh5.14.4.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,907 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - No path for the flink jar passed. Using the location of class 
> org.apache.flink.yarn.YarnClusterDescriptor to locate the jar
> 2019-12-18 15:04:41,084 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Cluster specification: 
> ClusterSpecification{masterMemoryMB=1024, taskManagerMemoryMB=12288, 
> numberTaskManagers=1, slotsPerTaskManager=6}
> 2019-12-18 15:04:42,344 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Submitting application master application_1576573857638_0026
> 2019-12-18 15:04:42,370 INFO  
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted 
> application application_1576573857638_0026
> 2019-12-18 15:04:42,371 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Waiting for the cluster to be allocated
> 2019-12-18 15:04:42,372 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Deploying cluster, current state ACCEPTED
> 2019-12-18 15:04:45,388 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - YARN application has been deployed successfully.
> 2019-12-18 15:04:45,390 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Found Web Interface debugboxcreate431x3.sa:36162 of 
> application 'application_1576573857638_0026'.
> Job has been submitted with JobID 9140c70769f4271cc22ea8becaa26272
> 
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: org.apache.flink.client.program.ProgramInvocationException: 
> Job failed (JobID: 9140c70769f4271cc22ea8becaa26272)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>   at 
> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>   at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:664)
>   at 

[jira] [Commented] (FLINK-15308) Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1

2019-12-18 Thread Feng Jiajie (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999706#comment-16999706
 ] 

Feng Jiajie commented on FLINK-15308:
-

Hi [~kevin.cyj] ,

I can reproduce the problem every time.

YARN cluster:  3 node ( 8 core 32GB )
{code:java}
$ cat flink-conf.yaml | grep -v '^#' | grep -v '^$'
jobmanager.rpc.address: localhost
jobmanager.rpc.port: 6123
jobmanager.heap.size: 1024m
taskmanager.memory.total-process.size: 1024m
taskmanager.numberOfTaskSlots: 6
parallelism.default: 1
taskmanager.network.pipelined-shuffle.compression.enabled: true
jobmanager.execution.failover-strategy: region
{code}

> Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1
> --
>
> Key: FLINK-15308
> URL: https://issues.apache.org/jira/browse/FLINK-15308
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.0
> Environment: $ git log
> commit 4b54da2c67692b1c9d43e1184c00899b0151b3ae
> Author: bowen.li 
> Date: Tue Dec 17 17:37:03 2019 -0800
>Reporter: Feng Jiajie
>Assignee: Yingjie Cao
>Priority: Blocker
> Attachments: image-2019-12-19-10-55-30-644.png
>
>
> Job worked well with default flink-conf.yaml with 
> pipelined-shuffle.compression:
> {code:java}
> taskmanager.numberOfTaskSlots: 1
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> But when I set taskmanager.numberOfTaskSlots to 4 or 6:
> {code:java}
> taskmanager.numberOfTaskSlots: 6
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> job failed:
> {code:java}
> $ bin/flink run -m yarn-cluster -p 16 -yjm 1024m -ytm 12288m 
> ~/flink-example-1.0-SNAPSHOT.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/lib/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/data/sa_cluster/cloudera/parcels/CDH-5.14.4-1.cdh5.14.4.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,907 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - No path for the flink jar passed. Using the location of class 
> org.apache.flink.yarn.YarnClusterDescriptor to locate the jar
> 2019-12-18 15:04:41,084 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Cluster specification: 
> ClusterSpecification{masterMemoryMB=1024, taskManagerMemoryMB=12288, 
> numberTaskManagers=1, slotsPerTaskManager=6}
> 2019-12-18 15:04:42,344 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Submitting application master application_1576573857638_0026
> 2019-12-18 15:04:42,370 INFO  
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted 
> application application_1576573857638_0026
> 2019-12-18 15:04:42,371 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Waiting for the cluster to be allocated
> 2019-12-18 15:04:42,372 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Deploying cluster, current state ACCEPTED
> 2019-12-18 15:04:45,388 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - YARN application has been deployed successfully.
> 2019-12-18 15:04:45,390 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Found Web Interface debugboxcreate431x3.sa:36162 of 
> application 'application_1576573857638_0026'.
> Job has been submitted with JobID 9140c70769f4271cc22ea8becaa26272
> 
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: org.apache.flink.client.program.ProgramInvocationException: 
> Job failed (JobID: 

[jira] [Issue Comment Deleted] (FLINK-15308) Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1

2019-12-18 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15308?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao updated FLINK-15308:

Comment: was deleted

(was: [~fengjiajie] I can not reproduce the problem in my test environment. Is 
there any other settings?

!image-2019-12-19-10-55-30-644.png!)

> Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1
> --
>
> Key: FLINK-15308
> URL: https://issues.apache.org/jira/browse/FLINK-15308
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.0
> Environment: $ git log
> commit 4b54da2c67692b1c9d43e1184c00899b0151b3ae
> Author: bowen.li 
> Date: Tue Dec 17 17:37:03 2019 -0800
>Reporter: Feng Jiajie
>Assignee: Yingjie Cao
>Priority: Blocker
> Attachments: image-2019-12-19-10-55-30-644.png
>
>
> Job worked well with default flink-conf.yaml with 
> pipelined-shuffle.compression:
> {code:java}
> taskmanager.numberOfTaskSlots: 1
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> But when I set taskmanager.numberOfTaskSlots to 4 or 6:
> {code:java}
> taskmanager.numberOfTaskSlots: 6
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> job failed:
> {code:java}
> $ bin/flink run -m yarn-cluster -p 16 -yjm 1024m -ytm 12288m 
> ~/flink-example-1.0-SNAPSHOT.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/lib/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/data/sa_cluster/cloudera/parcels/CDH-5.14.4-1.cdh5.14.4.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,907 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - No path for the flink jar passed. Using the location of class 
> org.apache.flink.yarn.YarnClusterDescriptor to locate the jar
> 2019-12-18 15:04:41,084 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Cluster specification: 
> ClusterSpecification{masterMemoryMB=1024, taskManagerMemoryMB=12288, 
> numberTaskManagers=1, slotsPerTaskManager=6}
> 2019-12-18 15:04:42,344 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Submitting application master application_1576573857638_0026
> 2019-12-18 15:04:42,370 INFO  
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted 
> application application_1576573857638_0026
> 2019-12-18 15:04:42,371 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Waiting for the cluster to be allocated
> 2019-12-18 15:04:42,372 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Deploying cluster, current state ACCEPTED
> 2019-12-18 15:04:45,388 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - YARN application has been deployed successfully.
> 2019-12-18 15:04:45,390 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Found Web Interface debugboxcreate431x3.sa:36162 of 
> application 'application_1576573857638_0026'.
> Job has been submitted with JobID 9140c70769f4271cc22ea8becaa26272
> 
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: org.apache.flink.client.program.ProgramInvocationException: 
> Job failed (JobID: 9140c70769f4271cc22ea8becaa26272)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>   at 
> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>   at 
> 

[jira] [Updated] (FLINK-15308) Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1

2019-12-18 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15308?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao updated FLINK-15308:

Attachment: image-2019-12-19-10-55-30-644.png

> Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1
> --
>
> Key: FLINK-15308
> URL: https://issues.apache.org/jira/browse/FLINK-15308
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.0
> Environment: $ git log
> commit 4b54da2c67692b1c9d43e1184c00899b0151b3ae
> Author: bowen.li 
> Date: Tue Dec 17 17:37:03 2019 -0800
>Reporter: Feng Jiajie
>Assignee: Yingjie Cao
>Priority: Blocker
> Attachments: image-2019-12-19-10-55-30-644.png
>
>
> Job worked well with default flink-conf.yaml with 
> pipelined-shuffle.compression:
> {code:java}
> taskmanager.numberOfTaskSlots: 1
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> But when I set taskmanager.numberOfTaskSlots to 4 or 6:
> {code:java}
> taskmanager.numberOfTaskSlots: 6
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> job failed:
> {code:java}
> $ bin/flink run -m yarn-cluster -p 16 -yjm 1024m -ytm 12288m 
> ~/flink-example-1.0-SNAPSHOT.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/lib/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/data/sa_cluster/cloudera/parcels/CDH-5.14.4-1.cdh5.14.4.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,907 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - No path for the flink jar passed. Using the location of class 
> org.apache.flink.yarn.YarnClusterDescriptor to locate the jar
> 2019-12-18 15:04:41,084 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Cluster specification: 
> ClusterSpecification{masterMemoryMB=1024, taskManagerMemoryMB=12288, 
> numberTaskManagers=1, slotsPerTaskManager=6}
> 2019-12-18 15:04:42,344 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Submitting application master application_1576573857638_0026
> 2019-12-18 15:04:42,370 INFO  
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted 
> application application_1576573857638_0026
> 2019-12-18 15:04:42,371 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Waiting for the cluster to be allocated
> 2019-12-18 15:04:42,372 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Deploying cluster, current state ACCEPTED
> 2019-12-18 15:04:45,388 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - YARN application has been deployed successfully.
> 2019-12-18 15:04:45,390 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Found Web Interface debugboxcreate431x3.sa:36162 of 
> application 'application_1576573857638_0026'.
> Job has been submitted with JobID 9140c70769f4271cc22ea8becaa26272
> 
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: org.apache.flink.client.program.ProgramInvocationException: 
> Job failed (JobID: 9140c70769f4271cc22ea8becaa26272)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>   at 
> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>   at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:664)
>   at 

[jira] [Commented] (FLINK-15308) Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1

2019-12-18 Thread Yingjie Cao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999694#comment-16999694
 ] 

Yingjie Cao commented on FLINK-15308:
-

[~fengjiajie] I can not reproduce the problem in my test environment. Is there 
any other settings?

!image-2019-12-19-10-55-30-644.png!

> Job failed when enable pipelined-shuffle.compression and numberOfTaskSlots > 1
> --
>
> Key: FLINK-15308
> URL: https://issues.apache.org/jira/browse/FLINK-15308
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.10.0
> Environment: $ git log
> commit 4b54da2c67692b1c9d43e1184c00899b0151b3ae
> Author: bowen.li 
> Date: Tue Dec 17 17:37:03 2019 -0800
>Reporter: Feng Jiajie
>Assignee: Yingjie Cao
>Priority: Blocker
> Attachments: image-2019-12-19-10-55-30-644.png
>
>
> Job worked well with default flink-conf.yaml with 
> pipelined-shuffle.compression:
> {code:java}
> taskmanager.numberOfTaskSlots: 1
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> But when I set taskmanager.numberOfTaskSlots to 4 or 6:
> {code:java}
> taskmanager.numberOfTaskSlots: 6
> taskmanager.network.pipelined-shuffle.compression.enabled: true
> {code}
> job failed:
> {code:java}
> $ bin/flink run -m yarn-cluster -p 16 -yjm 1024m -ytm 12288m 
> ~/flink-example-1.0-SNAPSHOT.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/lib/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/data/sa_cluster/cloudera/parcels/CDH-5.14.4-1.cdh5.14.4.p0.3/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,514 WARN  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>   - The configuration directory 
> ('/data/build/flink/flink-dist/target/flink-1.10-SNAPSHOT-bin/flink-1.10-SNAPSHOT/conf')
>  already contains a LOG4J config file.If you want to use logback, then please 
> delete or rename the log configuration file.
> 2019-12-18 15:04:40,907 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - No path for the flink jar passed. Using the location of class 
> org.apache.flink.yarn.YarnClusterDescriptor to locate the jar
> 2019-12-18 15:04:41,084 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Cluster specification: 
> ClusterSpecification{masterMemoryMB=1024, taskManagerMemoryMB=12288, 
> numberTaskManagers=1, slotsPerTaskManager=6}
> 2019-12-18 15:04:42,344 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Submitting application master application_1576573857638_0026
> 2019-12-18 15:04:42,370 INFO  
> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl - Submitted 
> application application_1576573857638_0026
> 2019-12-18 15:04:42,371 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Waiting for the cluster to be allocated
> 2019-12-18 15:04:42,372 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Deploying cluster, current state ACCEPTED
> 2019-12-18 15:04:45,388 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - YARN application has been deployed successfully.
> 2019-12-18 15:04:45,390 INFO  org.apache.flink.yarn.YarnClusterDescriptor 
>   - Found Web Interface debugboxcreate431x3.sa:36162 of 
> application 'application_1576573857638_0026'.
> Job has been submitted with JobID 9140c70769f4271cc22ea8becaa26272
> 
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: org.apache.flink.client.program.ProgramInvocationException: 
> Job failed (JobID: 9140c70769f4271cc22ea8becaa26272)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>   at 
> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>   at 

[jira] [Updated] (FLINK-15320) JobManager crashes in the standalone model when cancelling job which subtask' status is scheduled

2019-12-18 Thread lining (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lining updated FLINK-15320:
---
Summary: JobManager crashes in the standalone model when cancelling job 
which subtask' status is scheduled  (was: JobManager crash in the model of 
standalone)

> JobManager crashes in the standalone model when cancelling job which subtask' 
> status is scheduled
> -
>
> Key: FLINK-15320
> URL: https://issues.apache.org/jira/browse/FLINK-15320
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.10.0
>Reporter: lining
>Priority: Blocker
>
> Use start-cluster.sh to start a standalone cluster, and then submit a job 
> from the streaming's example which name is TopSpeedWindowing, parallelism is 
> 20. Wait for one minute, cancel the job, jobmanager will crash. The exception 
> stack is:
> 2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
> Could not assign resource 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
> current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
> TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print 
> to Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
>  at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
>  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) 
> at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.releaseSharedSlot(SchedulerImpl.java:552)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.cancelSlotRequest(SchedulerImpl.java:184)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.returnLogicalSlot(SchedulerImpl.java:195)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.lambda$returnSlotToOwner$0(SingleLogicalSlot.java:181)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> 

[jira] [Updated] (FLINK-15312) Remove PlanExposingEnvironment

2019-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-15312:
---
Labels: pull-request-available  (was: )

> Remove PlanExposingEnvironment
> --
>
> Key: FLINK-15312
> URL: https://issues.apache.org/jira/browse/FLINK-15312
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataSet, Tests
>Reporter: Zili Chen
>Assignee: Zili Chen
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> We don't need it. {{env.createProgramPlan}} is sufficient, which is internal 
> and can be used in testing.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] TisonKun opened a new pull request #10624: [FLINK-15312][tests] Remove PlanExposingEnvironment

2019-12-18 Thread GitBox
TisonKun opened a new pull request #10624: [FLINK-15312][tests] Remove 
PlanExposingEnvironment
URL: https://github.com/apache/flink/pull/10624
 
 
   ## What is the purpose of the change
   
   PlanExposingEnvironment is for testing only and we can replace it with 
`env.createProgramPlan` or an encapsulation of 
`PackagedProgramUtils#createPipeline`. We are towards a unique env entrypoint, 
I think.
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15320) JobManager crash in the model of standalone

2019-12-18 Thread lining (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lining updated FLINK-15320:
---
Priority: Blocker  (was: Critical)

> JobManager crash in the model of standalone
> ---
>
> Key: FLINK-15320
> URL: https://issues.apache.org/jira/browse/FLINK-15320
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.10.0
>Reporter: lining
>Priority: Blocker
>
> Use start-cluster.sh to start a standalone cluster, and then submit a job 
> from the streaming's example which name is TopSpeedWindowing, parallelism is 
> 20. Wait for one minute, cancel the job, jobmanager will crash. The exception 
> stack is:
> 2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...2019-12-19 10:12:11,060 ERROR 
> org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
> 'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
> Stopping the process...java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
> Could not assign resource 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
> current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
> TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print 
> to Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
>  at 
> org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
>  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) 
> at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.releaseSharedSlot(SchedulerImpl.java:552)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.cancelSlotRequest(SchedulerImpl.java:184)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.returnLogicalSlot(SchedulerImpl.java:195)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.lambda$returnSlotToOwner$0(SingleLogicalSlot.java:181)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenCompleteStage(CompletableFuture.java:778)
>  at 
> java.util.concurrent.CompletableFuture.whenComplete(CompletableFuture.java:2140)
>  at 
> org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.returnSlotToOwner(SingleLogicalSlot.java:178)
>  at 
> 

[jira] [Updated] (FLINK-15320) JobManager crash in the model of standalone

2019-12-18 Thread lining (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lining updated FLINK-15320:
---
Description: 
Use start-cluster.sh to start a standalone cluster, and then submit a job from 
the streaming's example which name is TopSpeedWindowing, parallelism is 20. 
Wait for one minute, cancel the job, jobmanager will crash. The exception stack 
is:

2019-12-19 10:12:11,060 ERROR 
org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
Stopping the process...2019-12-19 10:12:11,060 ERROR 
org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
Stopping the process...java.util.concurrent.CompletionException: 
java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
Could not assign resource 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print to 
Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
 at 
org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
 at 
java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) at 
java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474) 
at 
java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
 at 
org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
 at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474) 
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
 at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474) 
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
at 
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.releaseSharedSlot(SchedulerImpl.java:552)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.cancelSlotRequest(SchedulerImpl.java:184)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.returnLogicalSlot(SchedulerImpl.java:195)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.lambda$returnSlotToOwner$0(SingleLogicalSlot.java:181)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
 at 
java.util.concurrent.CompletableFuture.uniWhenCompleteStage(CompletableFuture.java:778)
 at 
java.util.concurrent.CompletableFuture.whenComplete(CompletableFuture.java:2140)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.returnSlotToOwner(SingleLogicalSlot.java:178)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.releaseSlot(SingleLogicalSlot.java:125)
 at 
org.apache.flink.runtime.executiongraph.Execution.releaseAssignedResource(Execution.java:1451)
 at 
org.apache.flink.runtime.executiongraph.Execution.finishCancellation(Execution.java:1170)
 at 
org.apache.flink.runtime.executiongraph.Execution.completeCancelling(Execution.java:1150)
 at 
org.apache.flink.runtime.executiongraph.Execution.completeCancelling(Execution.java:1129)
 at 
org.apache.flink.runtime.executiongraph.Execution.cancelAtomically(Execution.java:)
 at 

[jira] [Created] (FLINK-15320) JobManager crash in the model of standalone

2019-12-18 Thread lining (Jira)
lining created FLINK-15320:
--

 Summary: JobManager crash in the model of standalone
 Key: FLINK-15320
 URL: https://issues.apache.org/jira/browse/FLINK-15320
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.10.0
Reporter: lining


Use start-cluster.sh to start an independent cluster, and then submit a job 
from the stream example. The job name is TopSpeedWindowing, set concurrent. 20. 
Wait for a while, cancel the job, jobmanager will crash. The exception stack is:

2019-12-19 10:12:11,060 ERROR 
org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
Stopping the process...2019-12-19 10:12:11,060 ERROR 
org.apache.flink.runtime.util.FatalExitExceptionHandler       - FATAL: Thread 
'flink-akka.actor.default-dispatcher-2' produced an uncaught exception. 
Stopping the process...java.util.concurrent.CompletionException: 
java.util.concurrent.CompletionException: java.lang.IllegalStateException: 
Could not assign resource 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@583585c6 to 
current execution Attempt #0 (Window(GlobalWindows(), DeltaTrigger, 
TimeEvictor, ComparableAggregator, PassThroughWindowFunction) -> Sink: Print to 
Std. Out (19/20)) @ (unassigned) - [CANCELED]. at 
org.apache.flink.runtime.scheduler.DefaultScheduler.propagateIfNonNull(DefaultScheduler.java:387)
 at 
org.apache.flink.runtime.scheduler.DefaultScheduler.lambda$deployAll$4(DefaultScheduler.java:372)
 at 
java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822) at 
java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474) 
at 
java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
 at 
org.apache.flink.runtime.concurrent.FutureUtils$WaitingConjunctFuture.handleCompletedFuture(FutureUtils.java:705)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
 at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474) 
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.lambda$internalAllocateSlot$0(SchedulerImpl.java:170)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
 at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474) 
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962) 
at 
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.tryFulfillSlotRequestOrMakeAvailable(SlotPoolImpl.java:534)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSingleSlot(SlotPoolImpl.java:479)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.releaseSlot(SlotPoolImpl.java:390)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.release(SlotSharingManager.java:557)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.releaseChild(SlotSharingManager.java:607)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$MultiTaskSlot.access$700(SlotSharingManager.java:352)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SlotSharingManager$SingleTaskSlot.release(SlotSharingManager.java:716)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.releaseSharedSlot(SchedulerImpl.java:552)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.cancelSlotRequest(SchedulerImpl.java:184)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SchedulerImpl.returnLogicalSlot(SchedulerImpl.java:195)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.lambda$returnSlotToOwner$0(SingleLogicalSlot.java:181)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
 at 
java.util.concurrent.CompletableFuture.uniWhenCompleteStage(CompletableFuture.java:778)
 at 
java.util.concurrent.CompletableFuture.whenComplete(CompletableFuture.java:2140)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.returnSlotToOwner(SingleLogicalSlot.java:178)
 at 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot.releaseSlot(SingleLogicalSlot.java:125)
 at 
org.apache.flink.runtime.executiongraph.Execution.releaseAssignedResource(Execution.java:1451)
 at 
org.apache.flink.runtime.executiongraph.Execution.finishCancellation(Execution.java:1170)
 at 
org.apache.flink.runtime.executiongraph.Execution.completeCancelling(Execution.java:1150)
 at 

[jira] [Commented] (FLINK-10114) Support Orc for StreamingFileSink

2019-12-18 Thread vinoyang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-10114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999669#comment-16999669
 ] 

vinoyang commented on FLINK-10114:
--

Hi [~mbellani] thanks for paying attention to this feature and sharing your 
thought and design. I am sorry I have no time to join you and learn the 
context. And I am sorry I can not reassign this ticket. Based on the new 
contribution constitution, before implementing, it would be better to agree on 
the design with the community. In addition, it would be better to find a 
committer as your mentor for the feature. Ping [~aljoscha] and 
[~trohrm...@apache.org] to give more professional options.

> Support Orc for StreamingFileSink
> -
>
> Key: FLINK-10114
> URL: https://issues.apache.org/jira/browse/FLINK-10114
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / FileSystem
>Reporter: zhangminglei
>Assignee: vinoyang
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14172) Implement KubeClient with official Java client library for kubernetes

2019-12-18 Thread Jian He (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14172?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999625#comment-16999625
 ] 

Jian He commented on FLINK-14172:
-

I also see fabric8 client is used more widely than the official client 

> Implement KubeClient with official Java client library for kubernetes
> -
>
> Key: FLINK-14172
> URL: https://issues.apache.org/jira/browse/FLINK-14172
> Project: Flink
>  Issue Type: Sub-task
>  Components: Deployment / Kubernetes
>Reporter: Yang Wang
>Priority: Major
>
> Official Java client library for kubernetes is become more and more active. 
> The new features(such as leader election) and some client 
> implementations(informer, lister, cache) are better. So we should use the 
> official java client for kubernetes in flink.
> https://github.com/kubernetes-client/java



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-12160) Support to ship user config files on kubernetes

2019-12-18 Thread Jian He (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-12160?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999624#comment-16999624
 ] 

Jian He commented on FLINK-12160:
-

As the link noted, the  configMap is not suitable for large files like jars, 
reusing hdfs might be a better idea.

> Support to ship user config files on kubernetes
> ---
>
> Key: FLINK-12160
> URL: https://issues.apache.org/jira/browse/FLINK-12160
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Coordination
>Reporter: Yang Wang
>Assignee: Yang Wang
>Priority: Major
>
> When deploying flink cluster on Yarn, `yarnShip` cli option could be used to 
> ship user config files and jars. The files are registered as Yarn local 
> resource which is saved on hdfs.
>  
> However, we do not have a default existing dfs. The `ConfigMap` could be used 
> to ship small files. The files will be stored in etcd and then mounted to 
> jobmanager and taskmanager pod so that they could use it as locally.
> {code:java}
> -kt,--kubernetesship  Ship files in the specified 
> directory
>   (t for transfer). Only small 
> files(<1MB) could be supported.
> {code}
> [https://stackoverflow.com/questions/53012798/kubernetes-configmap-size-limitation]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15256) unable to drop table in HiveCatalogITCase

2019-12-18 Thread Bowen Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bowen Li closed FLINK-15256.

Resolution: Invalid

> unable to drop table in HiveCatalogITCase
> -
>
> Key: FLINK-15256
> URL: https://issues.apache.org/jira/browse/FLINK-15256
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Assignee: Terry Wang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> {code:java}
> @Test
>   public void testCsvTableViaSQL() throws Exception {
>   EnvironmentSettings settings = 
> EnvironmentSettings.newInstance().useBlinkPlanner().inBatchMode().build();
>   TableEnvironment tableEnv = TableEnvironment.create(settings);
>   tableEnv.registerCatalog("myhive", hiveCatalog);
>   tableEnv.useCatalog("myhive");
>   String path = 
> this.getClass().getResource("/csv/test.csv").getPath();
>   tableEnv.sqlUpdate("create table test2 (name String, age Int) 
> with (\n" +
>   "   'connector.type' = 'filesystem',\n" +
>   "   'connector.path' = 'file://" + path + "',\n" +
>   "   'format.type' = 'csv'\n" +
>   ")");
>   Table t = tableEnv.sqlQuery("SELECT * FROM 
> myhive.`default`.test2");
>   List result = TableUtils.collectToList(t);
>   // assert query result
>   assertEquals(
>   new HashSet<>(Arrays.asList(
>   Row.of("1", 1),
>   Row.of("2", 2),
>   Row.of("3", 3))),
>   new HashSet<>(result)
>   );
>   tableEnv.sqlUpdate("drop table myhive.`default`.tests2");
>   }
> {code}
> The last drop table statement reports error as:
> {code:java}
> org.apache.flink.table.api.ValidationException: Could not execute DropTable 
> in path `myhive`.`default`.`tests2`
>   at 
> org.apache.flink.table.catalog.CatalogManager.execute(CatalogManager.java:568)
>   at 
> org.apache.flink.table.catalog.CatalogManager.dropTable(CatalogManager.java:543)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:519)
>   at 
> org.apache.flink.table.catalog.hive.HiveCatalogITCase.testCsvTableViaSQL(HiveCatalogITCase.java:123)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>   at 
> org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169)
>   at 
> org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154)
>   at 
> org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
>   at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>   at 
> 

[jira] [Commented] (FLINK-15256) unable to drop table in HiveCatalogITCase

2019-12-18 Thread Bowen Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15256?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999470#comment-16999470
 ] 

Bowen Li commented on FLINK-15256:
--

[~Terry1897] I ran it today and it succeeded, maybe my local testing env having 
issues. Thanks for looking into it.

[~gjy] the merged PR has a wrong jira id.

Closing this as invalid

> unable to drop table in HiveCatalogITCase
> -
>
> Key: FLINK-15256
> URL: https://issues.apache.org/jira/browse/FLINK-15256
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Assignee: Terry Wang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> {code:java}
> @Test
>   public void testCsvTableViaSQL() throws Exception {
>   EnvironmentSettings settings = 
> EnvironmentSettings.newInstance().useBlinkPlanner().inBatchMode().build();
>   TableEnvironment tableEnv = TableEnvironment.create(settings);
>   tableEnv.registerCatalog("myhive", hiveCatalog);
>   tableEnv.useCatalog("myhive");
>   String path = 
> this.getClass().getResource("/csv/test.csv").getPath();
>   tableEnv.sqlUpdate("create table test2 (name String, age Int) 
> with (\n" +
>   "   'connector.type' = 'filesystem',\n" +
>   "   'connector.path' = 'file://" + path + "',\n" +
>   "   'format.type' = 'csv'\n" +
>   ")");
>   Table t = tableEnv.sqlQuery("SELECT * FROM 
> myhive.`default`.test2");
>   List result = TableUtils.collectToList(t);
>   // assert query result
>   assertEquals(
>   new HashSet<>(Arrays.asList(
>   Row.of("1", 1),
>   Row.of("2", 2),
>   Row.of("3", 3))),
>   new HashSet<>(result)
>   );
>   tableEnv.sqlUpdate("drop table myhive.`default`.tests2");
>   }
> {code}
> The last drop table statement reports error as:
> {code:java}
> org.apache.flink.table.api.ValidationException: Could not execute DropTable 
> in path `myhive`.`default`.`tests2`
>   at 
> org.apache.flink.table.catalog.CatalogManager.execute(CatalogManager.java:568)
>   at 
> org.apache.flink.table.catalog.CatalogManager.dropTable(CatalogManager.java:543)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:519)
>   at 
> org.apache.flink.table.catalog.hive.HiveCatalogITCase.testCsvTableViaSQL(HiveCatalogITCase.java:123)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>   at 
> org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runTestMethod(FlinkStandaloneHiveRunner.java:169)
>   at 
> org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:154)
>   at 
> org.apache.flink.connectors.hive.FlinkStandaloneHiveRunner.runChild(FlinkStandaloneHiveRunner.java:92)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>

[jira] [Assigned] (FLINK-15247) Closing (Testing)MiniCluster may cause ConcurrentModificationException

2019-12-18 Thread Andrey Zagrebin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15247?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrey Zagrebin reassigned FLINK-15247:
---

Assignee: Andrey Zagrebin

> Closing (Testing)MiniCluster may cause ConcurrentModificationException
> --
>
> Key: FLINK-15247
> URL: https://issues.apache.org/jira/browse/FLINK-15247
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Assignee: Andrey Zagrebin
>Priority: Blocker
>  Labels: test-stability
> Fix For: 1.10.0
>
>
> {noformat}
> Test 
> operatorsBecomeBackPressured(org.apache.flink.test.streaming.runtime.BackPressureITCase)
>  failed with:
> org.apache.flink.util.FlinkException: Could not close resource.
> at 
> org.apache.flink.util.AutoCloseableAsync.close(AutoCloseableAsync.java:42)org.apache.flink.test.streaming.runtime.BackPressureITCase.tearDown(BackPressureITCase.java:165)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33)
> at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at org.junit.runners.Suite.runChild(Suite.java:128)
> at org.junit.runners.Suite.runChild(Suite.java:27)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at 
> org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:107)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:83)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158)
> at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
> at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
> at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
> at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
> Caused by: org.apache.flink.util.FlinkException: Error while shutting the 
> TaskExecutor down.
> at 
> org.apache.flink.runtime.taskexecutor.TaskExecutor.handleOnStopException(TaskExecutor.java:397)
> at 
> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$onStop$0(TaskExecutor.java:382)
> at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:822)
> at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:797)
> at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
> at 
> 

[jira] [Commented] (FLINK-15199) Benchmarks are not compiling

2019-12-18 Thread Piotr Nowojski (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999369#comment-16999369
 ] 

Piotr Nowojski commented on FLINK-15199:


It's already merged to release-1.10 as 535845153ba45dc8e564f235e353e86aa1429fdd.

 

I probably forgot to specify it in my previous comment. 

> Benchmarks are not compiling
> 
>
> Key: FLINK-15199
> URL: https://issues.apache.org/jira/browse/FLINK-15199
> Project: Flink
>  Issue Type: Bug
>  Components: Benchmarks
>Affects Versions: 1.10.0
>Reporter: Piotr Nowojski
>Assignee: Piotr Nowojski
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Recent changes in FLINK-14926 caused:
> {noformat}
> [ERROR] COMPILATION ERROR : 
> [INFO] -
> [ERROR] 
> /home/jenkins/workspace/flink-master-benchmarks/flink-benchmarks/src/main/java/org/apache/flink/state/benchmark/BackendUtils.java:[61,57]
>  cannot infer type arguments for 
> org.apache.flink.contrib.streaming.state.RocksDBKeyedStateBackendBuilder<>
> [INFO] 1 error
> [INFO] -
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> [INFO] Total time: 2.635 s
> [INFO] Finished at: 2019-12-11T14:58:37+01:00
> [INFO] Final Memory: 31M/751M
> [INFO] 
> 
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) 
> on project flink-hackathon-benchmarks: Compilation failure
> [ERROR] 
> /home/jenkins/workspace/flink-master-benchmarks/flink-benchmarks/src/main/java/org/apache/flink/state/benchmark/BackendUtils.java:[61,57]
>  cannot infer type arguments for 
> org.apache.flink.contrib.streaming.state.RocksDBKeyedStateBackendBuilder<>
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-10114) Support Orc for StreamingFileSink

2019-12-18 Thread Manish Bellani (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-10114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999359#comment-16999359
 ] 

Manish Bellani commented on FLINK-10114:


Hi,

Is there still interest in this work? I have an implementation of an S3/Orc 
sink that handles these requirements. We're currently using this at GitHub 
internally to write several billions of events/terabytes of data per day to 
S3(exactly once). If there's interest I can kick off a conversation at GitHub, 
in the meantime do you mind sharing how can I contribute to flink? I found 
this: [https://flink.apache.org/contributing/contribute-code.html] but is there 
anything else that I need to do?

 

Thanks

Manish

> Support Orc for StreamingFileSink
> -
>
> Key: FLINK-10114
> URL: https://issues.apache.org/jira/browse/FLINK-10114
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / FileSystem
>Reporter: zhangminglei
>Assignee: vinoyang
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Issue Comment Deleted] (FLINK-9749) Rework Bucketing Sink

2019-12-18 Thread Manish Bellani (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-9749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Manish Bellani updated FLINK-9749:
--
Comment: was deleted

(was: Hi,

Is there still interest in this work? I have an implementation of an S3/Orc 
sink that handles these requirements. We're currently using this at GitHub 
internally to write several billions of events/terabytes of data per day to 
S3(exactly once). If there's interest I can kick off a conversation at GitHub, 
in the meantime do you mind sharing how can I contribute to flink? I found 
this: [https://flink.apache.org/contributing/contribute-code.html] but is there 
anything else that I need to do?

 

Thanks

Manish)

> Rework Bucketing Sink
> -
>
> Key: FLINK-9749
> URL: https://issues.apache.org/jira/browse/FLINK-9749
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / FileSystem
>Reporter: Stephan Ewen
>Assignee: Kostas Kloudas
>Priority: Major
>
> The BucketingSink has a series of deficits at the moment.
> Due to the long list of issues, I would suggest to add a new 
> StreamingFileSink with a new and cleaner design
> h3. Encoders, Parquet, ORC
>  - It only efficiently supports row-wise data formats (avro, jso, sequence 
> files.
>  - Efforts to add (columnar) compression for blocks of data is inefficient, 
> because blocks cannot span checkpoints due to persistence-on-checkpoint.
>  - The encoders are part of the \{{flink-connector-filesystem project}}, 
> rather than in orthogonal formats projects. This blows up the dependencies of 
> the \{{flink-connector-filesystem project}} project. As an example, the 
> rolling file sink has dependencies on Hadoop and Avro, which messes up 
> dependency management.
> h3. Use of FileSystems
>  - The BucketingSink works only on Hadoop's FileSystem abstraction not 
> support Flink's own FileSystem abstraction and cannot work with the packaged 
> S3, maprfs, and swift file systems
>  - The sink hence needs Hadoop as a dependency
>  - The sink relies on "trying out" whether truncation works, which requires 
> write access to the users working directory
>  - The sink relies on enumerating and counting files, rather than maintaining 
> its own state, making less efficient
> h3. Correctness and Efficiency on S3
>  - The BucketingSink relies on strong consistency in the file enumeration, 
> hence may work incorrectly on S3.
>  - The BucketingSink relies on persisting streams at intermediate points. 
> This is not working properly on S3, hence there may be data loss on S3.
> h3. .valid-length companion file
>  - The valid length file makes it hard for consumers of the data and should 
> be dropped
> We track this design in a series of sub issues.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-11956) Remove shading from S3 filesystems build

2019-12-18 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-11956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski closed FLINK-11956.
--
Fix Version/s: 1.10.0
   Resolution: Fixed

merged to master as 9fdedf552822d27efbd36f87d21b9b0f2637b757

merged to release-1.10 as 8ec545d56f007645ca8f2a2374386882132ffc7a

> Remove shading from S3 filesystems build
> 
>
> Key: FLINK-11956
> URL: https://issues.apache.org/jira/browse/FLINK-11956
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / FileSystem
>Affects Versions: 1.9.0
>Reporter: Stefan Richter
>Assignee: Arvid Heise
>Priority: Critical
> Fix For: 1.10.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-11956) Remove shading from S3 filesystems build

2019-12-18 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-11956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski updated FLINK-11956:
---
Summary: Remove shading from S3 filesystems build  (was: Remove shading 
from filesystems build)

> Remove shading from S3 filesystems build
> 
>
> Key: FLINK-11956
> URL: https://issues.apache.org/jira/browse/FLINK-11956
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / FileSystem
>Affects Versions: 1.9.0
>Reporter: Stefan Richter
>Assignee: Arvid Heise
>Priority: Critical
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (FLINK-15241) Revert the unexpected change of the configuration for Mesos CPU cores

2019-12-18 Thread Till Rohrmann (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15241?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Till Rohrmann resolved FLINK-15241.
---
Resolution: Fixed

Fixed via

master: 4fc55c747d001b75ba4652e867c98a5e2d62fd69
1.10.0: b360019433f25395a8fc151b23d94eb7bb81359d

> Revert the unexpected change of the configuration for Mesos CPU cores
> -
>
> Key: FLINK-15241
> URL: https://issues.apache.org/jira/browse/FLINK-15241
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Mesos
>Reporter: Yangze Guo
>Assignee: Yangze Guo
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Before 1.10, the default config of cpu cores for Mesos Workers is equal to 
> taskmanager.numberOfTaskSlots. However, it is now 0.0 by default. I feel 
> sorry I've not catch it in 
> [FLINK-15196|https://issues.apache.org/jira/browse/FLINK-15196]. I think this 
> change is unexpected to user thus we need to fix it.
> cc [~trohrmann] Could you assigned this ticket to me?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15199) Benchmarks are not compiling

2019-12-18 Thread Yu Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999320#comment-16999320
 ] 

Yu Li commented on FLINK-15199:
---

Waiting for this one to be merged also into release-1.10

> Benchmarks are not compiling
> 
>
> Key: FLINK-15199
> URL: https://issues.apache.org/jira/browse/FLINK-15199
> Project: Flink
>  Issue Type: Bug
>  Components: Benchmarks
>Affects Versions: 1.10.0
>Reporter: Piotr Nowojski
>Assignee: Piotr Nowojski
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Recent changes in FLINK-14926 caused:
> {noformat}
> [ERROR] COMPILATION ERROR : 
> [INFO] -
> [ERROR] 
> /home/jenkins/workspace/flink-master-benchmarks/flink-benchmarks/src/main/java/org/apache/flink/state/benchmark/BackendUtils.java:[61,57]
>  cannot infer type arguments for 
> org.apache.flink.contrib.streaming.state.RocksDBKeyedStateBackendBuilder<>
> [INFO] 1 error
> [INFO] -
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> [INFO] Total time: 2.635 s
> [INFO] Finished at: 2019-12-11T14:58:37+01:00
> [INFO] Final Memory: 31M/751M
> [INFO] 
> 
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.1:compile (default-compile) 
> on project flink-hackathon-benchmarks: Compilation failure
> [ERROR] 
> /home/jenkins/workspace/flink-master-benchmarks/flink-benchmarks/src/main/java/org/apache/flink/state/benchmark/BackendUtils.java:[61,57]
>  cannot infer type arguments for 
> org.apache.flink.contrib.streaming.state.RocksDBKeyedStateBackendBuilder<>
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (FLINK-14926) Make sure no resource leak of RocksObject

2019-12-18 Thread Yu Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14926?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yu Li resolved FLINK-14926.
---
Fix Version/s: 1.10.0
   Resolution: Fixed

Fixed in master via:
 * e05618b427
 * 0602ccf1d2
 * 93deb1a709
 * 31dd8c3624
 * 2405111bd6
 * 07f08fb19a

Fixed in 1.10.0 via:
 * 13c577f6d2
 * 5326730225
 * 386dad8705
 * f5fa21b7d0
 * 6f777d5a51
 * e94bf2fe70

(Thanks [~sewen] for reviewing and merging the PR)

> Make sure no resource leak of RocksObject
> -
>
> Key: FLINK-14926
> URL: https://issues.apache.org/jira/browse/FLINK-14926
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / State Backends
>Affects Versions: 1.6.4, 1.7.2, 1.8.2, 1.9.1
>Reporter: Yu Li
>Assignee: Yu Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> When investigating FLINK-14484 to allow setting {{WriteBufferManager}} via 
> options, we find it necessary to supply a {{close}} method in 
> {{OptionsFactory}} to make sure resources configured in options like 
> {{WriteBufferManager}} could/would be released.
> What's more, we're having potential risk of resource leak in 
> {{PredefinedOptions}} such as the created [BloomFilter 
> instance|https://github.com/apache/flink/blob/e14bd50fed42e89674ba9d01a231d5c5e59f490c/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/PredefinedOptions.java#L155]
>  is never closed (regression caused by FLINK-7220 changes), and we should 
> also fix this and add a {{close}} method in {{PredefinedOptions}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14926) Make sure no resource leak of RocksObject

2019-12-18 Thread Yu Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999315#comment-16999315
 ] 

Yu Li commented on FLINK-14926:
---

We also need to merge the changes of FLINK-15199 into release-1.10 to prevent 
compilation issue when running flink-benchmark against release-1.10, and I 
guess the plan was to include all related commits here when closing it. But yes 
I think we could also update FLINK-15199 to indicate the merging into 
release-1.10. 

Let me add some comments about changes for this JIRA and close it. Thanks for 
the reminder [~gjy].

> Make sure no resource leak of RocksObject
> -
>
> Key: FLINK-14926
> URL: https://issues.apache.org/jira/browse/FLINK-14926
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / State Backends
>Affects Versions: 1.6.4, 1.7.2, 1.8.2, 1.9.1
>Reporter: Yu Li
>Assignee: Yu Li
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> When investigating FLINK-14484 to allow setting {{WriteBufferManager}} via 
> options, we find it necessary to supply a {{close}} method in 
> {{OptionsFactory}} to make sure resources configured in options like 
> {{WriteBufferManager}} could/would be released.
> What's more, we're having potential risk of resource leak in 
> {{PredefinedOptions}} such as the created [BloomFilter 
> instance|https://github.com/apache/flink/blob/e14bd50fed42e89674ba9d01a231d5c5e59f490c/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/PredefinedOptions.java#L155]
>  is never closed (regression caused by FLINK-7220 changes), and we should 
> also fix this and add a {{close}} method in {{PredefinedOptions}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15317) State TTL Heap backend end-to-end test fails on Travis

2019-12-18 Thread Yu Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15317?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999314#comment-16999314
 ] 

Yu Li commented on FLINK-15317:
---

Thanks for logging the issue [~gjy] and the quick fix [~arvid heise].

Correct me if I'm wrong, is this a regression of FLINK-15076? If so, let's add 
some link to the issue. Thanks.

> State TTL Heap backend end-to-end test fails on Travis
> --
>
> Key: FLINK-15317
> URL: https://issues.apache.org/jira/browse/FLINK-15317
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network, Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Assignee: Arvid Heise
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> https://api.travis-ci.org/v3/job/626286529/log.txt
> {noformat}
> Checking for errors...
> Found error in log files:
> ...
> java.lang.IllegalStateException: Released
>   at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGate.getNextBufferOrEvent(SingleInputGate.java:483)
>   at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGate.pollNext(SingleInputGate.java:474)
>   at 
> org.apache.flink.runtime.taskmanager.InputGateWithMetrics.pollNext(InputGateWithMetrics.java:75)
>   at 
> org.apache.flink.streaming.runtime.io.CheckpointedInputGate.pollNext(CheckpointedInputGate.java:125)
>   at 
> org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:133)
>   at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:69)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:311)
>   at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:187)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:488)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:470)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:702)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:527)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13417) Bump Zookeeper to 3.5.5

2019-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13417?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-13417:
---
Labels: pull-request-available  (was: )

> Bump Zookeeper to 3.5.5
> ---
>
> Key: FLINK-13417
> URL: https://issues.apache.org/jira/browse/FLINK-13417
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0
>Reporter: Konstantin Knauf
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> User might want to secure their Zookeeper connection via SSL.
> This requires a Zookeeper version >= 3.5.1. We might as well try to bump it 
> to 3.5.5, which is the latest version. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] tillrohrmann opened a new pull request #10623: [WIP][FLINK-13417] Bump ZooKeeper from 3.4.10 to 3.5.6

2019-12-18 Thread GitBox
tillrohrmann opened a new pull request #10623: [WIP][FLINK-13417] Bump 
ZooKeeper from 3.4.10 to 3.5.6
URL: https://github.com/apache/flink/pull/10623
 
 
   This is a WIP PR for the ZooKeeper version upgrade.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-14364) Allow comments fail when not ignore parse errors in CsvRowDeserializationSchema

2019-12-18 Thread Jiayi Liao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14364?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999291#comment-16999291
 ] 

Jiayi Liao commented on FLINK-14364:


[~lzljs3620320] [~ykt836] 

Sorry for the late reply, I was busy on something else recently :(. 

I've added a validation in {{CsvValidator}} to improve the usability in this 
[PR|https://github.com/apache/flink/pull/10622]. I'll appreciate it if you can 
spare time and take a look.

> Allow comments fail when not ignore parse errors in 
> CsvRowDeserializationSchema
> ---
>
> Key: FLINK-14364
> URL: https://issues.apache.org/jira/browse/FLINK-14364
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Jingsong Lee
>Assignee: Jiayi Liao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Use CsvRowDeserializationSchema, when setIgnoreParseErrors(false) and 
> setAllowComments(true).
> If there are some comments in msg, will throw MismatchedInputException.
> If this a bug? and we should catch MismatchedInputException and return null?
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-13417) Bump Zookeeper to 3.5.5

2019-12-18 Thread Till Rohrmann (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999287#comment-16999287
 ] 

Till Rohrmann commented on FLINK-13417:
---

For future reference: 
[Here|https://lists.apache.org/thread.html/be402a11bc986219eabd9dd8af507f36f49784d5400d0873e9ec0c2e%40%3Cdev.flink.apache.org%3E]
 is the link to the dev ML discussion about dropping vendor specific 
repositories.

> Bump Zookeeper to 3.5.5
> ---
>
> Key: FLINK-13417
> URL: https://issues.apache.org/jira/browse/FLINK-13417
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0
>Reporter: Konstantin Knauf
>Priority: Major
> Fix For: 1.11.0
>
>
> User might want to secure their Zookeeper connection via SSL.
> This requires a Zookeeper version >= 3.5.1. We might as well try to bump it 
> to 3.5.5, which is the latest version. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14364) Allow comments fail when not ignore parse errors in CsvRowDeserializationSchema

2019-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-14364:
---
Labels: pull-request-available  (was: )

> Allow comments fail when not ignore parse errors in 
> CsvRowDeserializationSchema
> ---
>
> Key: FLINK-14364
> URL: https://issues.apache.org/jira/browse/FLINK-14364
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Jingsong Lee
>Assignee: Jiayi Liao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Use CsvRowDeserializationSchema, when setIgnoreParseErrors(false) and 
> setAllowComments(true).
> If there are some comments in msg, will throw MismatchedInputException.
> If this a bug? and we should catch MismatchedInputException and return null?
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] buptljy opened a new pull request #10622: [FLINK-14364] [flink-formats] Add validation when format.allow-comments is used with format.ignore-parse-errors in CsvValidator

2019-12-18 Thread GitBox
buptljy opened a new pull request #10622: [FLINK-14364] [flink-formats] Add 
validation when format.allow-comments is used with format.ignore-parse-errors 
in CsvValidator
URL: https://github.com/apache/flink/pull/10622
 
 
   
   ## What is the purpose of the change
   
   As described in 
[FLINK-14364](https://issues.apache.org/jira/browse/FLINK-14364), we can only 
use `format.allow-comments` when  `format.ignore-parse-errors` is set. So we 
add a validation here to improve the usability.
   
   
   ## Brief change log
   
   Add a new check in `CsvValidator`.
   
   
   ## Verifying this change
   
   Unit testing.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (FLINK-13417) Bump Zookeeper to 3.5.5

2019-12-18 Thread Zili Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999279#comment-16999279
 ] 

Zili Chen edited comment on FLINK-13417 at 12/18/19 3:40 PM:
-

Thanks for reviving this thread [~sewen] & [~trohrmann]. One thing I'd like to 
emphasize again is that the bump to zk 3.5 will force user update version of 
their zookeeper ensemble. See also [this 
comment|https://issues.apache.org/jira/browse/FLINK-13417?focusedCommentId=16921386=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16921386]


was (Author: tison):
Thanks for reviving this thread [~sewen][~trohrmann]. One thing I'd like to 
emphasize again is that the bump to zk 3.5 will force user update version of 
their zookeeper ensemble. See also [this 
comment|https://issues.apache.org/jira/browse/FLINK-13417?focusedCommentId=16921386=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16921386]

> Bump Zookeeper to 3.5.5
> ---
>
> Key: FLINK-13417
> URL: https://issues.apache.org/jira/browse/FLINK-13417
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0
>Reporter: Konstantin Knauf
>Priority: Major
> Fix For: 1.11.0
>
>
> User might want to secure their Zookeeper connection via SSL.
> This requires a Zookeeper version >= 3.5.1. We might as well try to bump it 
> to 3.5.5, which is the latest version. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-13417) Bump Zookeeper to 3.5.5

2019-12-18 Thread Zili Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999279#comment-16999279
 ] 

Zili Chen commented on FLINK-13417:
---

Thanks for reviving this thread [~sewen][~trohrmann]. One thing I'd like to 
emphasize again is that the bump to zk 3.5 will force user update version of 
their zookeeper ensemble. See also [this 
comment|https://issues.apache.org/jira/browse/FLINK-13417?focusedCommentId=16921386=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16921386]

> Bump Zookeeper to 3.5.5
> ---
>
> Key: FLINK-13417
> URL: https://issues.apache.org/jira/browse/FLINK-13417
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0
>Reporter: Konstantin Knauf
>Priority: Major
> Fix For: 1.11.0
>
>
> User might want to secure their Zookeeper connection via SSL.
> This requires a Zookeeper version >= 3.5.1. We might as well try to bump it 
> to 3.5.5, which is the latest version. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-13417) Bump Zookeeper to 3.5.5

2019-12-18 Thread Till Rohrmann (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999263#comment-16999263
 ] 

Till Rohrmann commented on FLINK-13417:
---

Hi [~tison], thanks a lot for driving this effort. I would like to help with 
finalizing it so that we can include it in the upcoming 1.10 release. As a 
first step I would propose to drop the vendor specific repos. I'll start a 
discussion on dev for this. Then the version upgrade should be a bit easier.

> Bump Zookeeper to 3.5.5
> ---
>
> Key: FLINK-13417
> URL: https://issues.apache.org/jira/browse/FLINK-13417
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0
>Reporter: Konstantin Knauf
>Priority: Major
> Fix For: 1.11.0
>
>
> User might want to secure their Zookeeper connection via SSL.
> This requires a Zookeeper version >= 3.5.1. We might as well try to bump it 
> to 3.5.5, which is the latest version. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15247) Closing (Testing)MiniCluster may cause ConcurrentModificationException

2019-12-18 Thread Andrey Zagrebin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15247?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999249#comment-16999249
 ] 

Andrey Zagrebin commented on FLINK-15247:
-

It seems the problem is a bit deeper.

We introduced the delayed TaskExecutor#stopTaskExecutorServices in case of 
stopping TM in FLINK-11630 to gracefully cancel the tasks and let them release 
the resources of services before the final shutdown of services. Previously, 
slots were static, there were no resources attached to slots and TaskSlotTable 
was stateless, that did not cause issues.

Now TaskSlotTable has dynamically allocated slots after 
[FLINK-14189|https://jira.apache.org/jira/browse/FLINK-14189] and memory 
manager is attached to the slot after 
[FLINK-14400|https://jira.apache.org/jira/browse/FLINK-14400]. It means that 
when we cancel now tasks it triggers stopping of the services (including 
TaskSlotTable) concurrently with freeing slots and their resources. Even we put 
stopping of the services into the main thread, freeing of slots can be after 
stopping. It would be cleaner to firstly cancel tasks, wait for the subsequent 
freeing of the slots and then shutdown services.

cc [~trohrm...@apache.org]

> Closing (Testing)MiniCluster may cause ConcurrentModificationException
> --
>
> Key: FLINK-15247
> URL: https://issues.apache.org/jira/browse/FLINK-15247
> Project: Flink
>  Issue Type: Bug
>  Components: Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Priority: Blocker
>  Labels: test-stability
> Fix For: 1.10.0
>
>
> {noformat}
> Test 
> operatorsBecomeBackPressured(org.apache.flink.test.streaming.runtime.BackPressureITCase)
>  failed with:
> org.apache.flink.util.FlinkException: Could not close resource.
> at 
> org.apache.flink.util.AutoCloseableAsync.close(AutoCloseableAsync.java:42)org.apache.flink.test.streaming.runtime.BackPressureITCase.tearDown(BackPressureITCase.java:165)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33)
> at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at org.junit.runners.Suite.runChild(Suite.java:128)
> at org.junit.runners.Suite.runChild(Suite.java:27)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at 
> org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:107)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:83)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75)
> at 
> org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158)
> at 
> 

[jira] [Updated] (FLINK-15317) State TTL Heap backend end-to-end test fails on Travis

2019-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15317?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-15317:
---
Labels: pull-request-available  (was: )

> State TTL Heap backend end-to-end test fails on Travis
> --
>
> Key: FLINK-15317
> URL: https://issues.apache.org/jira/browse/FLINK-15317
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network, Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Assignee: Arvid Heise
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>
> https://api.travis-ci.org/v3/job/626286529/log.txt
> {noformat}
> Checking for errors...
> Found error in log files:
> ...
> java.lang.IllegalStateException: Released
>   at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGate.getNextBufferOrEvent(SingleInputGate.java:483)
>   at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGate.pollNext(SingleInputGate.java:474)
>   at 
> org.apache.flink.runtime.taskmanager.InputGateWithMetrics.pollNext(InputGateWithMetrics.java:75)
>   at 
> org.apache.flink.streaming.runtime.io.CheckpointedInputGate.pollNext(CheckpointedInputGate.java:125)
>   at 
> org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:133)
>   at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:69)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:311)
>   at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:187)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:488)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:470)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:702)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:527)
>   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] AHeise opened a new pull request #10621: [FLINK-15317][runtime] Handle closed input gates more gracefully during

2019-12-18 Thread GitBox
AHeise opened a new pull request #10621: [FLINK-15317][runtime] Handle closed 
input gates more gracefully during
URL: https://github.com/apache/flink/pull/10621
 
 
   cancellation.
   
   
   
   ## What is the purpose of the change
   
   During task cancellation, input gates can be closed in the task cancellation 
thread, before the main invokable has been interrupted, which resulted in reads 
from a closed gate.
   
   This commit will handle this case in a more graceful manner such that the 
end-user is not seeing any additional exception.
   
   ## Brief change log
   
   - Use `TaskCancelException` instead of `IllegalStateException` when input 
gates are closed in a similar fashion as this exception has been used in 
`LocalInputChannel`.
   - Add a special handling of `TaskCancelException` to 
`StreamTask#runMainLoop()`, to make sure the exception is not appearing in the 
log.
   
   ## Verifying this change
   
   This fix is indirectly covered by many e2e tests that rely on this exception 
being swallowed.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / **no** / don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-15244) FileUtils#deleteDirectoryQuietly will delete files in the symbolic link which point to a directory

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao closed FLINK-15244.

Resolution: Fixed

1.10: 337f1cc534aac7f2806465687eeefb4d4984e3f3
master: 3c8a3ac0e7c5784a125d2bce3d2058fd5d1066a8

> FileUtils#deleteDirectoryQuietly will delete files in the symbolic link which 
> point to a directory
> --
>
> Key: FLINK-15244
> URL: https://issues.apache.org/jira/browse/FLINK-15244
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0
>Reporter: Wei Zhong
>Assignee: Wei Zhong
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> _FileUtils.deleteDirectoryQuietly_ will delete files in symbolic link which 
> point to a directory. Currently the PyFlink uses this method to delete 
> temporary folders generated during the job submission and python UDF 
> execution, which contains the symbolic links which may point to users' 
> libraries and directories in distributed cache. 
> To resolve this problem we need to check if the directory is symbolic link in 
> _FileUtils.deleteDirectoryInternal:_
> {code:java}
> private static void deleteDirectoryInternal(File directory) throws 
> IOException {
> // **We should check if the directory is symbolic link.**
> if (directory.isDirectory()) {
>   // directory exists and is a directory
>   // empty the directory first
>   try {
>  cleanDirectoryInternal(directory);
>   }
>   catch (FileNotFoundException ignored) {
>  // someone concurrently deleted the directory, nothing to do for us
>  return;
>   }
>   // delete the directory. this fails if the directory is not empty, 
> meaning
>   // if new files got concurrently created. we want to fail then.
>   // if someone else deleted the empty directory concurrently, we don't 
> mind
>   // the result is the same for us, after all
>   Files.deleteIfExists(directory.toPath());
>}
>else if (directory.exists()) {
>   // exists but is file, not directory
>   // either an error from the caller, or concurrently a file got created
>   throw new IOException(directory + " is not a directory");
>}
>// else: does not exist, which is okay (as if deleted)
> }
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15244) FileUtils#deleteDirectoryQuietly will delete files in the symbolic link which point to a directory

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15244:
-
Fix Version/s: (was: 1.11.0)

> FileUtils#deleteDirectoryQuietly will delete files in the symbolic link which 
> point to a directory
> --
>
> Key: FLINK-15244
> URL: https://issues.apache.org/jira/browse/FLINK-15244
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0
>Reporter: Wei Zhong
>Assignee: Wei Zhong
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> _FileUtils.deleteDirectoryQuietly_ will delete files in symbolic link which 
> point to a directory. Currently the PyFlink uses this method to delete 
> temporary folders generated during the job submission and python UDF 
> execution, which contains the symbolic links which may point to users' 
> libraries and directories in distributed cache. 
> To resolve this problem we need to check if the directory is symbolic link in 
> _FileUtils.deleteDirectoryInternal:_
> {code:java}
> private static void deleteDirectoryInternal(File directory) throws 
> IOException {
> // **We should check if the directory is symbolic link.**
> if (directory.isDirectory()) {
>   // directory exists and is a directory
>   // empty the directory first
>   try {
>  cleanDirectoryInternal(directory);
>   }
>   catch (FileNotFoundException ignored) {
>  // someone concurrently deleted the directory, nothing to do for us
>  return;
>   }
>   // delete the directory. this fails if the directory is not empty, 
> meaning
>   // if new files got concurrently created. we want to fail then.
>   // if someone else deleted the empty directory concurrently, we don't 
> mind
>   // the result is the same for us, after all
>   Files.deleteIfExists(directory.toPath());
>}
>else if (directory.exists()) {
>   // exists but is file, not directory
>   // either an error from the caller, or concurrently a file got created
>   throw new IOException(directory + " is not a directory");
>}
>// else: does not exist, which is okay (as if deleted)
> }
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15219) LocalEnvironment is not initializing plugins

2019-12-18 Thread Arvid Heise (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999166#comment-16999166
 ] 

Arvid Heise commented on FLINK-15219:
-

Piotr suggested to provide a manual hook for users to initialize 
`Plugins.initialize()` in their respective main. Then we can later support 
non-filesystem plugins.

> LocalEnvironment is not initializing plugins
> 
>
> Key: FLINK-15219
> URL: https://issues.apache.org/jira/browse/FLINK-15219
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.10.0
>Reporter: Arvid Heise
>Priority: Major
>
> Currently, plugins cannot be used in `LocalEnvironment` as `FileSystem`s are 
> only initialized without `PluginManager`. In all other environments, the 
> plugin manager is initialized in `TaskManagerRunner`.
>  
> That means unless a user manually adds
> {code:java}
>final Configuration configuration = new Configuration();
>FileSystem.initialize(configuration, 
> PluginUtils.createPluginManagerFromRootFolder(configuration));
> {code}
> to his entry point, he cannot use plugins, which means no s3 filesystems 
> after FLINK-11956.
> We should add the proper initialization to the LocalExecutionEnvironment (at 
> LocalExecutor?), so that the user does not have to it manually.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-13197) Fix Hive view row type mismatch when expanding in planner

2019-12-18 Thread Dawid Wysakowicz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Wysakowicz closed FLINK-13197.

Resolution: Fixed

Fixed in:
master: 4e4dcf521748538aa164124fd613c9fe096ea604
1.10: 3d70c75477705b71acc83c010e7b89fce6c78c19

> Fix Hive view row type mismatch when expanding in planner
> -
>
> Key: FLINK-13197
> URL: https://issues.apache.org/jira/browse/FLINK-13197
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive, Table SQL / Planner
>Reporter: Bowen Li
>Assignee: Rui Li
>Priority: Major
> Fix For: 1.10.0
>
>
> One goal of HiveCatalog and hive integration is to enable Flink-Hive 
> interoperability, that is Flink should understand existing Hive meta-objects, 
> and Hive meta-objects created thru Flink should be understood by Hive.
> Taking an example of a Hive view v1 in HiveCatalog and database hc.db. Unlike 
> an equivalent Flink view whose full path in expanded query should be 
> hc.db.v1, the Hive view's full path in the expanded query should be db.v1 
> such that Hive can understand it, no matter it's created by Hive or Flink.
> [~lirui] can you help to ensure that Flink can also query Hive's view in both 
> Flink planner and Blink planner?
> cc [~xuefuz]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13197) Fix Hive view row type mismatch when expanding in planner

2019-12-18 Thread Dawid Wysakowicz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Wysakowicz updated FLINK-13197:
-
Component/s: Table SQL / Planner

> Fix Hive view row type mismatch when expanding in planner
> -
>
> Key: FLINK-13197
> URL: https://issues.apache.org/jira/browse/FLINK-13197
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive, Table SQL / Planner
>Reporter: Bowen Li
>Assignee: Rui Li
>Priority: Major
> Fix For: 1.10.0
>
>
> One goal of HiveCatalog and hive integration is to enable Flink-Hive 
> interoperability, that is Flink should understand existing Hive meta-objects, 
> and Hive meta-objects created thru Flink should be understood by Hive.
> Taking an example of a Hive view v1 in HiveCatalog and database hc.db. Unlike 
> an equivalent Flink view whose full path in expanded query should be 
> hc.db.v1, the Hive view's full path in the expanded query should be db.v1 
> such that Hive can understand it, no matter it's created by Hive or Flink.
> [~lirui] can you help to ensure that Flink can also query Hive's view in both 
> Flink planner and Blink planner?
> cc [~xuefuz]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13197) Fix Hive view row type mismatch when expanding in planner

2019-12-18 Thread Dawid Wysakowicz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Wysakowicz updated FLINK-13197:
-
Issue Type: Bug  (was: Test)

> Fix Hive view row type mismatch when expanding in planner
> -
>
> Key: FLINK-13197
> URL: https://issues.apache.org/jira/browse/FLINK-13197
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Assignee: Rui Li
>Priority: Major
> Fix For: 1.10.0
>
>
> One goal of HiveCatalog and hive integration is to enable Flink-Hive 
> interoperability, that is Flink should understand existing Hive meta-objects, 
> and Hive meta-objects created thru Flink should be understood by Hive.
> Taking an example of a Hive view v1 in HiveCatalog and database hc.db. Unlike 
> an equivalent Flink view whose full path in expanded query should be 
> hc.db.v1, the Hive view's full path in the expanded query should be db.v1 
> such that Hive can understand it, no matter it's created by Hive or Flink.
> [~lirui] can you help to ensure that Flink can also query Hive's view in both 
> Flink planner and Blink planner?
> cc [~xuefuz]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15186) Column names and nullability mismatch when querying CatalogViews

2019-12-18 Thread Dawid Wysakowicz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15186?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Wysakowicz closed FLINK-15186.

Fix Version/s: (was: 1.10.0)
   Resolution: Duplicate

> Column names and nullability mismatch when querying CatalogViews
> 
>
> Key: FLINK-15186
> URL: https://issues.apache.org/jira/browse/FLINK-15186
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Legacy Planner, Table SQL / Planner
>Reporter: Rui Li
>Priority: Major
>
> When querying a CatalogView, the column names and nullability we get from the 
> catalog may differ from what we arrive at after compiling the view's query. 
> And we fail to expand the view due to this mismatch.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-15289) Run sql appear error of "Zero-length character strings have no serializable string representation".

2019-12-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-15289:
---

Assignee: Jingsong Lee

> Run sql appear error of "Zero-length character strings have no serializable 
> string representation".
> ---
>
> Key: FLINK-15289
> URL: https://issues.apache.org/jira/browse/FLINK-15289
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Assignee: Jingsong Lee
>Priority: Critical
> Fix For: 1.10.0
>
>
> *The sql is:*
>  CREATE TABLE `INT8_TBL` (
>  q1 BIGINT,
>  q2 BIGINT
>  ) WITH (
>  'format.field-delimiter'=',',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  
> 'connector.path'='/defender_test_data/daily_regression_batch_postgres_1.10/test_bigint/sources/INT8_TBL.csv',
>  'format.type'='csv'
>  );
> SELECT '' AS five, q1 AS plus, -q1 AS xm FROM INT8_TBL;
> *The error detail is :*
>  2019-12-17 15:35:07,026 ERROR org.apache.flink.table.client.SqlClient - SQL 
> Client must stop. Unexpected exception. This is a bug. Please consider filing 
> an issue.
>  org.apache.flink.table.api.TableException: Zero-length character strings 
> have no serializable string representation.
>  at 
> org.apache.flink.table.types.logical.CharType.asSerializableString(CharType.java:116)
>  at 
> org.apache.flink.table.descriptors.DescriptorProperties.putTableSchema(DescriptorProperties.java:218)
>  at 
> org.apache.flink.table.catalog.CatalogTableImpl.toProperties(CatalogTableImpl.java:75)
>  at 
> org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSink(TableFactoryUtil.java:85)
>  at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeQueryAndPersistInternal(LocalExecutor.java:688)
>  at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeQueryAndPersist(LocalExecutor.java:488)
>  at org.apache.flink.table.client.cli.CliClient.callSelect(CliClient.java:601)
>  at 
> org.apache.flink.table.client.cli.CliClient.callCommand(CliClient.java:385)
>  at java.util.Optional.ifPresent(Optional.java:159)
>  at 
> org.apache.flink.table.client.cli.CliClient.submitSQLFile(CliClient.java:271)
>  at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:125)
>  at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>  at org.apache.flink.table.client.SqlClient.main(SqlClient.java:180)
> *The input data is:*
>  123,456
>  123,4567890123456789
>  4567890123456789,123
>  4567890123456789,4567890123456789
>  4567890123456789,-4567890123456789



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-15313) Can not insert decimal with precision into sink using TypeInformation

2019-12-18 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15313?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-15313:
---

Assignee: Jark Wu

> Can not insert decimal with precision into sink using TypeInformation
> -
>
> Key: FLINK-15313
> URL: https://issues.apache.org/jira/browse/FLINK-15313
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Jingsong Lee
>Assignee: Jark Wu
>Priority: Blocker
> Fix For: 1.10.0
>
>
> Sink DDL:
>  
> {code:java}
> val sinkDDL =
>   s"""
> |CREATE TABLE T2 (
> |  d DECIMAL(10, 2),
> |  cnt INT
> |) with (
> |  'connector.type' = 'filesystem',
> |  'connector.path' = '$sinkFilePath',
> |  'format.type' = 'csv',
> |  'format.field-delimiter' = ','
> |)
>   """.stripMargin
> {code}
> Using blink with batch mode. (ensure insert BinaryRow into sink table)
>  
> In FLINK-15124 , but we still use wrong precision to construct 
> DataFormatConverter. This will lead to exception when inserting BinaryRow. 
> (BinaryRow need correct precision to get)
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15289) Run sql appear error of "Zero-length character strings have no serializable string representation".

2019-12-18 Thread Timo Walther (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999128#comment-16999128
 ] 

Timo Walther commented on FLINK-15289:
--

Yes, we should do that.

> Run sql appear error of "Zero-length character strings have no serializable 
> string representation".
> ---
>
> Key: FLINK-15289
> URL: https://issues.apache.org/jira/browse/FLINK-15289
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Critical
> Fix For: 1.10.0
>
>
> *The sql is:*
>  CREATE TABLE `INT8_TBL` (
>  q1 BIGINT,
>  q2 BIGINT
>  ) WITH (
>  'format.field-delimiter'=',',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  
> 'connector.path'='/defender_test_data/daily_regression_batch_postgres_1.10/test_bigint/sources/INT8_TBL.csv',
>  'format.type'='csv'
>  );
> SELECT '' AS five, q1 AS plus, -q1 AS xm FROM INT8_TBL;
> *The error detail is :*
>  2019-12-17 15:35:07,026 ERROR org.apache.flink.table.client.SqlClient - SQL 
> Client must stop. Unexpected exception. This is a bug. Please consider filing 
> an issue.
>  org.apache.flink.table.api.TableException: Zero-length character strings 
> have no serializable string representation.
>  at 
> org.apache.flink.table.types.logical.CharType.asSerializableString(CharType.java:116)
>  at 
> org.apache.flink.table.descriptors.DescriptorProperties.putTableSchema(DescriptorProperties.java:218)
>  at 
> org.apache.flink.table.catalog.CatalogTableImpl.toProperties(CatalogTableImpl.java:75)
>  at 
> org.apache.flink.table.factories.TableFactoryUtil.findAndCreateTableSink(TableFactoryUtil.java:85)
>  at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeQueryAndPersistInternal(LocalExecutor.java:688)
>  at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeQueryAndPersist(LocalExecutor.java:488)
>  at org.apache.flink.table.client.cli.CliClient.callSelect(CliClient.java:601)
>  at 
> org.apache.flink.table.client.cli.CliClient.callCommand(CliClient.java:385)
>  at java.util.Optional.ifPresent(Optional.java:159)
>  at 
> org.apache.flink.table.client.cli.CliClient.submitSQLFile(CliClient.java:271)
>  at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:125)
>  at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>  at org.apache.flink.table.client.SqlClient.main(SqlClient.java:180)
> *The input data is:*
>  123,456
>  123,4567890123456789
>  4567890123456789,123
>  4567890123456789,4567890123456789
>  4567890123456789,-4567890123456789



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15319) flink-end-to-end-tests-common-kafka fails due to timeout

2019-12-18 Thread Yun Tang (Jira)
Yun Tang created FLINK-15319:


 Summary: flink-end-to-end-tests-common-kafka fails due to timeout
 Key: FLINK-15319
 URL: https://issues.apache.org/jira/browse/FLINK-15319
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
Affects Versions: 1.11.0
Reporter: Yun Tang


{code:java}
17:47:50.589 [INFO] Running 
org.apache.flink.tests.util.kafka.StreamingKafkaITCase
17:51:35.273 [ERROR] Tests run: 3, Failures: 0, Errors: 1, Skipped: 0, Time 
elapsed: 224.68 s <<< FAILURE! - in 
org.apache.flink.tests.util.kafka.StreamingKafkaITCase
17:51:35.273 [ERROR] testKafka[2: 
kafka-version:2.2.0](org.apache.flink.tests.util.kafka.StreamingKafkaITCase)  
Time elapsed: 120.008 s  <<< ERROR!
java.io.IOException: Process failed due to timeout.

{code}
Instance: [https://api.travis-ci.com/v3/job/268159458/log.txt]

 

It seems logs not uploaded to transfer.sh



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15318) RocksDBWriteBatchPerformanceTest.benchMark fails on ppc64le

2019-12-18 Thread Yun Tang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999114#comment-16999114
 ] 

Yun Tang commented on FLINK-15318:
--

Which version of Flink did you verify? Did you observe a stable performance 
behavior before and then suddenly fail due to sync with new commits or you just 
come across this error for the first time when you just want to try it on 
ppc64le platform.

Actually, Flink community lacks of benchmark on ppc64le environment and I 
noticed that RocksDB on ppc64le behaves not as good as those on linux64.

> RocksDBWriteBatchPerformanceTest.benchMark fails on ppc64le
> ---
>
> Key: FLINK-15318
> URL: https://issues.apache.org/jira/browse/FLINK-15318
> Project: Flink
>  Issue Type: Bug
>  Components: Benchmarks, Runtime / State Backends
> Environment: arch: ppc64le
> os: rhel7.6, ubuntu 18.04
> jdk: 8, 11
> mvn: 3.3.9, 3.6.2
>Reporter: Siddhesh Ghadi
>Priority: Major
> Attachments: surefire-report.txt
>
>
> RocksDBWriteBatchPerformanceTest.benchMark fails due to TestTimedOut, however 
> when test-timeout is increased from 2s to 5s in 
> org/apache/flink/contrib/streaming/state/benchmark/RocksDBWriteBatchPerformanceTest.java:75,
>  it passes. Is this acceptable solution?
> Note: Tests are ran inside a container.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-14944) Unstable test FlinkFnExecutionSyncTests.test_flink_fn_execution_pb2_synced

2019-12-18 Thread Hequn Cheng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hequn Cheng closed FLINK-14944.
---
Resolution: Resolved

> Unstable test FlinkFnExecutionSyncTests.test_flink_fn_execution_pb2_synced
> --
>
> Key: FLINK-14944
> URL: https://issues.apache.org/jira/browse/FLINK-14944
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python
>Affects Versions: 1.10.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This tests failed occasionally:
> {code:java}
> force = 'True', output_dir = '/tmp/tmpkh6rmeig'
> def generate_proto_files(force=True, 
> output_dir=DEFAULT_PYTHON_OUTPUT_PATH):
> try:
> import grpc_tools  # noqa  # pylint: disable=unused-import
> except ImportError:
> warnings.warn('Installing grpcio-tools is recommended for 
> development.')
> 
> proto_dirs = [os.path.join(PYFLINK_ROOT_PATH, path) for path in 
> PROTO_PATHS]
> proto_files = sum(
> [glob.glob(os.path.join(d, '*.proto')) for d in proto_dirs], [])
> out_dir = os.path.join(PYFLINK_ROOT_PATH, output_dir)
> out_files = [path for path in glob.glob(os.path.join(out_dir, 
> '*_pb2.py'))]
> 
> if out_files and not proto_files and not force:
> # We have out_files but no protos; assume they're up to date.
> # This is actually the common case (e.g. installation from an 
> sdist).
> logging.info('No proto files; using existing generated files.')
> return
> 
> elif not out_files and not proto_files:
> raise RuntimeError(
> 'No proto files found in %s.' % proto_dirs)
> 
> # Regenerate iff the proto files or this file are newer.
> elif force or not out_files or len(out_files) < len(proto_files) or (
> min(os.path.getmtime(path) for path in out_files)
> <= max(os.path.getmtime(path)
>for path in proto_files + 
> [os.path.realpath(__file__)])):
> try:
> >   from grpc_tools import protoc
> E   ModuleNotFoundError: No module named 'grpc_tools'
> pyflink/gen_protos.py:70: ModuleNotFoundError
> During handling of the above exception, another exception occurred:
> self = 
>   testMethod=test_flink_fn_execution_pb2_synced>
> def test_flink_fn_execution_pb2_synced(self):
> >   generate_proto_files('True', self.tempdir)
> pyflink/fn_execution/tests/test_flink_fn_execution_pb2_synced.py:35: 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ 
> pyflink/gen_protos.py:83: in generate_proto_files
> target=_install_grpcio_tools_and_generate_proto_files(force, output_dir))
> pyflink/gen_protos.py:131: in _install_grpcio_tools_and_generate_proto_files
> '--upgrade', GRPC_TOOLS, "-I"])
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ 
> popenargs = 
> (['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', '-m', 
> 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  ...],)
> kwargs = {}, retcode = 2
> cmd = ['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', 
> '-m', 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  ...]
> def check_call(*popenargs, **kwargs):
> """Run command with arguments.  Wait for command to complete.  If
> the exit code was zero then return, otherwise raise
> CalledProcessError.  The CalledProcessError object will have the
> return code in the returncode attribute.
> 
> The arguments are the same as for the call function.  Example:
> 
> check_call(["ls", "-l"])
> """
> retcode = call(*popenargs, **kwargs)
> if retcode:
> cmd = kwargs.get("args")
> if cmd is None:
> cmd = popenargs[0]
> >   raise CalledProcessError(retcode, cmd)
> E   subprocess.CalledProcessError: Command 
> '['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', '-m', 
> 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  '--build', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels-build',
>  '--upgrade', 'grpcio-tools>=1.3.5,<=1.14.2', '-I']' returned non-zero exit 
> status 2.
> dev/.conda/envs/3.6/lib/python3.6/subprocess.py:311: CalledProcessError
>  Captured stdout setup 

[jira] [Updated] (FLINK-15268) Shaded Hadoop S3A end-to-end test fails on travis

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15268?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15268:
-
Fix Version/s: 1.10.0

> Shaded Hadoop S3A end-to-end test fails on travis
> -
>
> Key: FLINK-15268
> URL: https://issues.apache.org/jira/browse/FLINK-15268
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem, Tests
>Affects Versions: 1.10.0
>Reporter: Yu Li
>Priority: Blocker
> Fix For: 1.10.0
>
>
> As titled, the 'Shaded Hadoop S3A end-to-end test' case failed with below 
> error:
> {code}
> java.io.IOException: regular upload failed: java.lang.NoClassDefFoundError: 
> javax/xml/bind/JAXBException
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3AUtils.extractException(S3AUtils.java:291)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3ABlockOutputStream.putObject(S3ABlockOutputStream.java:448)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3ABlockOutputStream.close(S3ABlockOutputStream.java:360)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:72)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:101)
>   at 
> org.apache.flink.fs.s3.common.hadoop.HadoopDataOutputStream.close(HadoopDataOutputStream.java:52)
>   at 
> org.apache.flink.core.fs.ClosingFSDataOutputStream.close(ClosingFSDataOutputStream.java:64)
>   at 
> java.base/java.io.FilterOutputStream.close(FilterOutputStream.java:188)
>   at java.base/sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:341)
>   at java.base/sun.nio.cs.StreamEncoder.close(StreamEncoder.java:161)
>   at 
> java.base/java.io.OutputStreamWriter.close(OutputStreamWriter.java:258)
>   at 
> org.apache.flink.api.java.io.CsvOutputFormat.close(CsvOutputFormat.java:170)
>   at 
> org.apache.flink.runtime.operators.DataSinkTask.invoke(DataSinkTask.java:227)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:702)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:527)
>   at java.base/java.lang.Thread.run(Thread.java:834)
> Caused by: java.lang.NoClassDefFoundError: javax/xml/bind/JAXBException
>   at 
> org.apache.flink.fs.s3base.shaded.com.amazonaws.util.Md5Utils.md5AsBase64(Md5Utils.java:104)
>   at 
> org.apache.flink.fs.s3base.shaded.com.amazonaws.services.s3.AmazonS3Client.putObject(AmazonS3Client.java:1647)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3AFileSystem.putObjectDirect(S3AFileSystem.java:1531)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.WriteOperationHelper.lambda$putObject$5(WriteOperationHelper.java:426)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.Invoker.once(Invoker.java:109)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.Invoker.lambda$retry$3(Invoker.java:260)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:317)
> {code}
> https://api.travis-ci.org/v3/job/625037121/log.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15268) Shaded Hadoop S3A end-to-end test fails on travis

2019-12-18 Thread Arvid Heise (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15268?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999092#comment-16999092
 ] 

Arvid Heise commented on FLINK-15268:
-

Is this related to Java 11 not having jaxb anymore?

> Shaded Hadoop S3A end-to-end test fails on travis
> -
>
> Key: FLINK-15268
> URL: https://issues.apache.org/jira/browse/FLINK-15268
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem, Tests
>Affects Versions: 1.10.0
>Reporter: Yu Li
>Priority: Blocker
>
> As titled, the 'Shaded Hadoop S3A end-to-end test' case failed with below 
> error:
> {code}
> java.io.IOException: regular upload failed: java.lang.NoClassDefFoundError: 
> javax/xml/bind/JAXBException
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3AUtils.extractException(S3AUtils.java:291)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3ABlockOutputStream.putObject(S3ABlockOutputStream.java:448)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3ABlockOutputStream.close(S3ABlockOutputStream.java:360)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:72)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:101)
>   at 
> org.apache.flink.fs.s3.common.hadoop.HadoopDataOutputStream.close(HadoopDataOutputStream.java:52)
>   at 
> org.apache.flink.core.fs.ClosingFSDataOutputStream.close(ClosingFSDataOutputStream.java:64)
>   at 
> java.base/java.io.FilterOutputStream.close(FilterOutputStream.java:188)
>   at java.base/sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:341)
>   at java.base/sun.nio.cs.StreamEncoder.close(StreamEncoder.java:161)
>   at 
> java.base/java.io.OutputStreamWriter.close(OutputStreamWriter.java:258)
>   at 
> org.apache.flink.api.java.io.CsvOutputFormat.close(CsvOutputFormat.java:170)
>   at 
> org.apache.flink.runtime.operators.DataSinkTask.invoke(DataSinkTask.java:227)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:702)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:527)
>   at java.base/java.lang.Thread.run(Thread.java:834)
> Caused by: java.lang.NoClassDefFoundError: javax/xml/bind/JAXBException
>   at 
> org.apache.flink.fs.s3base.shaded.com.amazonaws.util.Md5Utils.md5AsBase64(Md5Utils.java:104)
>   at 
> org.apache.flink.fs.s3base.shaded.com.amazonaws.services.s3.AmazonS3Client.putObject(AmazonS3Client.java:1647)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.S3AFileSystem.putObjectDirect(S3AFileSystem.java:1531)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.WriteOperationHelper.lambda$putObject$5(WriteOperationHelper.java:426)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.Invoker.once(Invoker.java:109)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.Invoker.lambda$retry$3(Invoker.java:260)
>   at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:317)
> {code}
> https://api.travis-ci.org/v3/job/625037121/log.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-9009) Error| You are creating too many HashedWheelTimer instances. HashedWheelTimer is a shared resource that must be reused across the application, so that only a few

2019-12-18 Thread Subramanyam Ramanathan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-9009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999056#comment-16999056
 ] 

Subramanyam Ramanathan edited comment on FLINK-9009 at 12/18/19 11:52 AM:
--

Hi,

I'm seeing a similar issue when using flink with Pulsar source + Sink.

I am using flink 1.8.2 and pulsar v2.4.2, on an 8 cpu 16GB RAM vm, running 
centos 7

I have 20 map transformations, each with it's own source and sink, and 
parallelism set to 8.

If the source and sink are Kafka, then I don't see any error, and top command 
shows me 4% memory usage.

When I use pulsar source+sink, the java process consumes *40 %* memory which is 
around 6GB. This happens even if I have not streamed any data.

TaskManager heap size was set to 1024M and I don't see any outOfMemory errors. 
I think the increase in memory usage is because flink uses the off heap memory 
which gets set by flink to -XX:MaxDirectMemorySize=8388607T,and something with 
Pulsar Source/Sink is causing it to consume a lot of it.

I also see the message in the logs mentioned in the title : *"Error: You are 
creating too many HashedWheelTimer instances. HashedWheelTimer is a shared 
resource."* 

 

Can you please help me understand the behaviour of the off heap memory in this 
case, and why it grows so much?

Is there any fix that is planned for this ? Or any way I can work around this ? 

 


was (Author: subbu-ramanathan107):
Hi,

I'm seeing a similar issue when using flink with Pulsar source + Sink.

I am using flink 1.8.2 and pulsar v2.4.2, on an 8 cpu 16GB RAM vm, running 
centos 7

I have 20 map transformations, each with it's own source and sink, and 
parallelism set to 8.

If the source and sink are Kafka, then I don't see any error, and top command 
shows me 4% memory usage.

When I use pulsar source+sink, the java process consumes *40 %* memory. This 
happens even if I have not streamed any data.

TaskManager heap size was set to 1024M and I don't see any outOfMemory errors. 
I think the increase in memory usage is because flink uses the off heap memory 
which gets set by flink to -XX:MaxDirectMemorySize=8388607T,and something with 
Pulsar Source/Sink is causing it to consume a lot of it.

I also see the message in the logs mentioned in the title : *"Error: You are 
creating too many HashedWheelTimer instances. HashedWheelTimer is a shared 
resource."* 

 

Can you please help me understand the behaviour of the off heap memory in this 
case, and why it grows so much?

Is there any fix that is planned for this ? Or any way I can work around this ? 

 

> Error| You are creating too many HashedWheelTimer instances.  
> HashedWheelTimer is a shared resource that must be reused across the 
> application, so that only a few instances are created.
> -
>
> Key: FLINK-9009
> URL: https://issues.apache.org/jira/browse/FLINK-9009
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
> Environment: Pass platform: Openshift
>Reporter: Pankaj
>Priority: Major
>
> Steps to reproduce:
> 1- Flink with Kafka as a consumer -> Writing stream to Cassandra using flink 
> cassandra sink.
> 2- In memory Job manager and task manager with checkpointing 5000ms.
> 3- env.setpararllelism(10)-> As kafka topic has 10 partition.
> 4- There are around 13 unique streams in a single flink run time environment 
> which are reading from kafka -> processing and writing to cassandra.
> Hardware: CPU 200 milli core . It is deployed on Paas platform on one node
> Memory: 526 MB.
>  
> When i start the server, It starts flink and all off sudden stops with above 
> error. It also shows out of memory error.
>  
> It would be nice if any body can suggest if something is wrong.
>  
> Maven:
> flink-connector-cassandra_2.11: 1.3.2
> flink-streaming-java_2.11: 1.4.0
> flink-connector-kafka-0.11_2.11:1.4.0
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-9009) Error| You are creating too many HashedWheelTimer instances. HashedWheelTimer is a shared resource that must be reused across the application, so that only a few

2019-12-18 Thread Subramanyam Ramanathan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-9009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999056#comment-16999056
 ] 

Subramanyam Ramanathan edited comment on FLINK-9009 at 12/18/19 11:52 AM:
--

Hi,

I'm seeing a similar issue when using flink with Pulsar source + Sink.

I am using flink 1.8.2 and pulsar v2.4.2, on an 8 cpu 16GB RAM vm, running 
centos 7

I have 20 map transformations, each with it's own source and sink, and 
parallelism set to 8.

If the source and sink are Kafka, then I don't see any error, and top command 
shows me 4% memory usage.

When I use pulsar source+sink, the java process consumes *40 %* memory. This 
happens even if I have not streamed any data.

TaskManager heap size was set to 1024M and I don't see any outOfMemory errors. 
I think the increase in memory usage is because flink uses the off heap memory 
which gets set by flink to -XX:MaxDirectMemorySize=8388607T,and something with 
Pulsar Source/Sink is causing it to consume a lot of it.

I also see the message in the logs mentioned in the title : *"Error: You are 
creating too many HashedWheelTimer instances. HashedWheelTimer is a shared 
resource."* 

 

Can you please help me understand the behaviour of the off heap memory in this 
case, and why it grows so much?

Is there any fix that is planned for this ? Or any way I can work around this ? 

 


was (Author: subbu-ramanathan107):
Hi,

I'm seeing a similar issue when using flink with Pulsar source + Sink.

I am using flink 1.8.2 and pulsar v2.4.2, on an 8 cpu 16GB RAM vm, running 
centos 7

I have 20 map transformations, each with it's own source and sink, and 
parallelism set to 8.

If the source and sink are Kafka, then I don't see any error, and top command 
shows me 4% memory usage.

When I use pulsar source+sink, the java process consumes *40 %* memory. This 
happens even if I have not streamed any data.

My heap size was set to 1024M and I don't see any outOfMemory errors. I think 
the increase in memory usage is because flink uses the off heap memory which 
gets set by flink to -XX:MaxDirectMemorySize=8388607T,and something with Pulsar 
Source/Sink is causing it to consume a lot of it.

I also see the message in the logs mentioned in the title : *"Error: You are 
creating too many HashedWheelTimer instances. HashedWheelTimer is a shared 
resource."* 

 

Can you please help me understand the behaviour of the off heap memory in this 
case, and why it grows so much?

Is there any fix that is planned for this ? Or any way I can work around this ? 

 

> Error| You are creating too many HashedWheelTimer instances.  
> HashedWheelTimer is a shared resource that must be reused across the 
> application, so that only a few instances are created.
> -
>
> Key: FLINK-9009
> URL: https://issues.apache.org/jira/browse/FLINK-9009
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
> Environment: Pass platform: Openshift
>Reporter: Pankaj
>Priority: Major
>
> Steps to reproduce:
> 1- Flink with Kafka as a consumer -> Writing stream to Cassandra using flink 
> cassandra sink.
> 2- In memory Job manager and task manager with checkpointing 5000ms.
> 3- env.setpararllelism(10)-> As kafka topic has 10 partition.
> 4- There are around 13 unique streams in a single flink run time environment 
> which are reading from kafka -> processing and writing to cassandra.
> Hardware: CPU 200 milli core . It is deployed on Paas platform on one node
> Memory: 526 MB.
>  
> When i start the server, It starts flink and all off sudden stops with above 
> error. It also shows out of memory error.
>  
> It would be nice if any body can suggest if something is wrong.
>  
> Maven:
> flink-connector-cassandra_2.11: 1.3.2
> flink-streaming-java_2.11: 1.4.0
> flink-connector-kafka-0.11_2.11:1.4.0
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-9009) Error| You are creating too many HashedWheelTimer instances. HashedWheelTimer is a shared resource that must be reused across the application, so that only a few

2019-12-18 Thread Subramanyam Ramanathan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-9009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999056#comment-16999056
 ] 

Subramanyam Ramanathan edited comment on FLINK-9009 at 12/18/19 11:52 AM:
--

Hi,

I'm seeing a similar issue when using flink with Pulsar source + Sink.

I am using flink 1.8.2 and pulsar v2.4.2, on an 8 cpu 16GB RAM vm, running 
centos 7

I have 20 map transformations, each with it's own source and sink, and 
parallelism set to 8.

If the source and sink are Kafka, then I don't see any error, and top command 
shows me 4% memory usage.

When I use pulsar source+sink, the java process consumes *40 %* memory which is 
around 6GB. This happens even if I have not streamed any data.

TaskManager heap size was set to 1024M and I don't see any outOfMemory errors. 
I think the increase in memory usage is because flink uses the off heap memory 
which gets set by flink to -XX:MaxDirectMemorySize=8388607T,and something with 
Pulsar Source/Sink is causing it to consume a lot of it.

I also see the error message in the logs mentioned in the title : *"Error: You 
are creating too many HashedWheelTimer instances. HashedWheelTimer is a shared 
resource."* 

 

Can you please help me understand the behaviour of the off heap memory in this 
case, and why it grows so much?

Is there any fix that is planned for this ? Or any way I can work around this ? 

 


was (Author: subbu-ramanathan107):
Hi,

I'm seeing a similar issue when using flink with Pulsar source + Sink.

I am using flink 1.8.2 and pulsar v2.4.2, on an 8 cpu 16GB RAM vm, running 
centos 7

I have 20 map transformations, each with it's own source and sink, and 
parallelism set to 8.

If the source and sink are Kafka, then I don't see any error, and top command 
shows me 4% memory usage.

When I use pulsar source+sink, the java process consumes *40 %* memory which is 
around 6GB. This happens even if I have not streamed any data.

TaskManager heap size was set to 1024M and I don't see any outOfMemory errors. 
I think the increase in memory usage is because flink uses the off heap memory 
which gets set by flink to -XX:MaxDirectMemorySize=8388607T,and something with 
Pulsar Source/Sink is causing it to consume a lot of it.

I also see the message in the logs mentioned in the title : *"Error: You are 
creating too many HashedWheelTimer instances. HashedWheelTimer is a shared 
resource."* 

 

Can you please help me understand the behaviour of the off heap memory in this 
case, and why it grows so much?

Is there any fix that is planned for this ? Or any way I can work around this ? 

 

> Error| You are creating too many HashedWheelTimer instances.  
> HashedWheelTimer is a shared resource that must be reused across the 
> application, so that only a few instances are created.
> -
>
> Key: FLINK-9009
> URL: https://issues.apache.org/jira/browse/FLINK-9009
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
> Environment: Pass platform: Openshift
>Reporter: Pankaj
>Priority: Major
>
> Steps to reproduce:
> 1- Flink with Kafka as a consumer -> Writing stream to Cassandra using flink 
> cassandra sink.
> 2- In memory Job manager and task manager with checkpointing 5000ms.
> 3- env.setpararllelism(10)-> As kafka topic has 10 partition.
> 4- There are around 13 unique streams in a single flink run time environment 
> which are reading from kafka -> processing and writing to cassandra.
> Hardware: CPU 200 milli core . It is deployed on Paas platform on one node
> Memory: 526 MB.
>  
> When i start the server, It starts flink and all off sudden stops with above 
> error. It also shows out of memory error.
>  
> It would be nice if any body can suggest if something is wrong.
>  
> Maven:
> flink-connector-cassandra_2.11: 1.3.2
> flink-streaming-java_2.11: 1.4.0
> flink-connector-kafka-0.11_2.11:1.4.0
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14926) Make sure no resource leak of RocksObject

2019-12-18 Thread Gary Yao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999082#comment-16999082
 ] 

Gary Yao commented on FLINK-14926:
--

What's the state here? There are already commits in release-1.10, e.g., 
208de7c29f2ca71b526f98f603cf4a55103c5520.

> Make sure no resource leak of RocksObject
> -
>
> Key: FLINK-14926
> URL: https://issues.apache.org/jira/browse/FLINK-14926
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / State Backends
>Affects Versions: 1.6.4, 1.7.2, 1.8.2, 1.9.1
>Reporter: Yu Li
>Assignee: Yu Li
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> When investigating FLINK-14484 to allow setting {{WriteBufferManager}} via 
> options, we find it necessary to supply a {{close}} method in 
> {{OptionsFactory}} to make sure resources configured in options like 
> {{WriteBufferManager}} could/would be released.
> What's more, we're having potential risk of resource leak in 
> {{PredefinedOptions}} such as the created [BloomFilter 
> instance|https://github.com/apache/flink/blob/e14bd50fed42e89674ba9d01a231d5c5e59f490c/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/PredefinedOptions.java#L155]
>  is never closed (regression caused by FLINK-7220 changes), and we should 
> also fix this and add a {{close}} method in {{PredefinedOptions}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14944) Unstable test FlinkFnExecutionSyncTests.test_flink_fn_execution_pb2_synced

2019-12-18 Thread Hequn Cheng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hequn Cheng updated FLINK-14944:

Issue Type: Improvement  (was: Bug)

> Unstable test FlinkFnExecutionSyncTests.test_flink_fn_execution_pb2_synced
> --
>
> Key: FLINK-14944
> URL: https://issues.apache.org/jira/browse/FLINK-14944
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python
>Affects Versions: 1.10.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This tests failed occasionally:
> {code:java}
> force = 'True', output_dir = '/tmp/tmpkh6rmeig'
> def generate_proto_files(force=True, 
> output_dir=DEFAULT_PYTHON_OUTPUT_PATH):
> try:
> import grpc_tools  # noqa  # pylint: disable=unused-import
> except ImportError:
> warnings.warn('Installing grpcio-tools is recommended for 
> development.')
> 
> proto_dirs = [os.path.join(PYFLINK_ROOT_PATH, path) for path in 
> PROTO_PATHS]
> proto_files = sum(
> [glob.glob(os.path.join(d, '*.proto')) for d in proto_dirs], [])
> out_dir = os.path.join(PYFLINK_ROOT_PATH, output_dir)
> out_files = [path for path in glob.glob(os.path.join(out_dir, 
> '*_pb2.py'))]
> 
> if out_files and not proto_files and not force:
> # We have out_files but no protos; assume they're up to date.
> # This is actually the common case (e.g. installation from an 
> sdist).
> logging.info('No proto files; using existing generated files.')
> return
> 
> elif not out_files and not proto_files:
> raise RuntimeError(
> 'No proto files found in %s.' % proto_dirs)
> 
> # Regenerate iff the proto files or this file are newer.
> elif force or not out_files or len(out_files) < len(proto_files) or (
> min(os.path.getmtime(path) for path in out_files)
> <= max(os.path.getmtime(path)
>for path in proto_files + 
> [os.path.realpath(__file__)])):
> try:
> >   from grpc_tools import protoc
> E   ModuleNotFoundError: No module named 'grpc_tools'
> pyflink/gen_protos.py:70: ModuleNotFoundError
> During handling of the above exception, another exception occurred:
> self = 
>   testMethod=test_flink_fn_execution_pb2_synced>
> def test_flink_fn_execution_pb2_synced(self):
> >   generate_proto_files('True', self.tempdir)
> pyflink/fn_execution/tests/test_flink_fn_execution_pb2_synced.py:35: 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ 
> pyflink/gen_protos.py:83: in generate_proto_files
> target=_install_grpcio_tools_and_generate_proto_files(force, output_dir))
> pyflink/gen_protos.py:131: in _install_grpcio_tools_and_generate_proto_files
> '--upgrade', GRPC_TOOLS, "-I"])
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ 
> popenargs = 
> (['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', '-m', 
> 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  ...],)
> kwargs = {}, retcode = 2
> cmd = ['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', 
> '-m', 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  ...]
> def check_call(*popenargs, **kwargs):
> """Run command with arguments.  Wait for command to complete.  If
> the exit code was zero then return, otherwise raise
> CalledProcessError.  The CalledProcessError object will have the
> return code in the returncode attribute.
> 
> The arguments are the same as for the call function.  Example:
> 
> check_call(["ls", "-l"])
> """
> retcode = call(*popenargs, **kwargs)
> if retcode:
> cmd = kwargs.get("args")
> if cmd is None:
> cmd = popenargs[0]
> >   raise CalledProcessError(retcode, cmd)
> E   subprocess.CalledProcessError: Command 
> '['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', '-m', 
> 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  '--build', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels-build',
>  '--upgrade', 'grpcio-tools>=1.3.5,<=1.14.2', '-I']' returned non-zero exit 
> status 2.
> dev/.conda/envs/3.6/lib/python3.6/subprocess.py:311: CalledProcessError
>  

[jira] [Commented] (FLINK-14944) Unstable test FlinkFnExecutionSyncTests.test_flink_fn_execution_pb2_synced

2019-12-18 Thread Hequn Cheng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14944?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16999080#comment-16999080
 ] 

Hequn Cheng commented on FLINK-14944:
-

Resloved 
in 1.11.0 via a7884abde4646745a30f964181c10c9f04b03b4a
in 1.10.0 via 729e87cf3371d8b42383564d45caaf4518afc4c9

> Unstable test FlinkFnExecutionSyncTests.test_flink_fn_execution_pb2_synced
> --
>
> Key: FLINK-14944
> URL: https://issues.apache.org/jira/browse/FLINK-14944
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.10.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This tests failed occasionally:
> {code:java}
> force = 'True', output_dir = '/tmp/tmpkh6rmeig'
> def generate_proto_files(force=True, 
> output_dir=DEFAULT_PYTHON_OUTPUT_PATH):
> try:
> import grpc_tools  # noqa  # pylint: disable=unused-import
> except ImportError:
> warnings.warn('Installing grpcio-tools is recommended for 
> development.')
> 
> proto_dirs = [os.path.join(PYFLINK_ROOT_PATH, path) for path in 
> PROTO_PATHS]
> proto_files = sum(
> [glob.glob(os.path.join(d, '*.proto')) for d in proto_dirs], [])
> out_dir = os.path.join(PYFLINK_ROOT_PATH, output_dir)
> out_files = [path for path in glob.glob(os.path.join(out_dir, 
> '*_pb2.py'))]
> 
> if out_files and not proto_files and not force:
> # We have out_files but no protos; assume they're up to date.
> # This is actually the common case (e.g. installation from an 
> sdist).
> logging.info('No proto files; using existing generated files.')
> return
> 
> elif not out_files and not proto_files:
> raise RuntimeError(
> 'No proto files found in %s.' % proto_dirs)
> 
> # Regenerate iff the proto files or this file are newer.
> elif force or not out_files or len(out_files) < len(proto_files) or (
> min(os.path.getmtime(path) for path in out_files)
> <= max(os.path.getmtime(path)
>for path in proto_files + 
> [os.path.realpath(__file__)])):
> try:
> >   from grpc_tools import protoc
> E   ModuleNotFoundError: No module named 'grpc_tools'
> pyflink/gen_protos.py:70: ModuleNotFoundError
> During handling of the above exception, another exception occurred:
> self = 
>   testMethod=test_flink_fn_execution_pb2_synced>
> def test_flink_fn_execution_pb2_synced(self):
> >   generate_proto_files('True', self.tempdir)
> pyflink/fn_execution/tests/test_flink_fn_execution_pb2_synced.py:35: 
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ 
> pyflink/gen_protos.py:83: in generate_proto_files
> target=_install_grpcio_tools_and_generate_proto_files(force, output_dir))
> pyflink/gen_protos.py:131: in _install_grpcio_tools_and_generate_proto_files
> '--upgrade', GRPC_TOOLS, "-I"])
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
> _ 
> popenargs = 
> (['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', '-m', 
> 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  ...],)
> kwargs = {}, retcode = 2
> cmd = ['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', 
> '-m', 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  ...]
> def check_call(*popenargs, **kwargs):
> """Run command with arguments.  Wait for command to complete.  If
> the exit code was zero then return, otherwise raise
> CalledProcessError.  The CalledProcessError object will have the
> return code in the returncode attribute.
> 
> The arguments are the same as for the call function.  Example:
> 
> check_call(["ls", "-l"])
> """
> retcode = call(*popenargs, **kwargs)
> if retcode:
> cmd = kwargs.get("args")
> if cmd is None:
> cmd = popenargs[0]
> >   raise CalledProcessError(retcode, cmd)
> E   subprocess.CalledProcessError: Command 
> '['/home/travis/build/apache/flink/flink-python/.tox/py36/bin/python', '-m', 
> 'pip', 'install', '--prefix', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels',
>  '--build', 
> '/home/travis/build/apache/flink/flink-python/pyflink/../.eggs/grpcio-wheels-build',
>  '--upgrade', 'grpcio-tools>=1.3.5,<=1.14.2', '-I']' returned non-zero exit 
> status 

[jira] [Reopened] (FLINK-15316) SQL Client end-to-end test (Old planner) failed on Travis

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao reopened FLINK-15316:
--

> SQL Client end-to-end test (Old planner) failed on Travis
> -
>
> Key: FLINK-15316
> URL: https://issues.apache.org/jira/browse/FLINK-15316
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client, Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.10.0
>
>
> https://api.travis-ci.org/v3/job/626286501/log.txt
> {noformat}
> Checking for errors...
> Found error in log files:
> ...
> org.apache.flink.table.client.gateway.SqlExecutionException: Invalid SQL 
> update statement.
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.applyUpdate(LocalExecutor.java:697)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdateInternal(LocalExecutor.java:576)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdate(LocalExecutor.java:527)
>   at 
> org.apache.flink.table.client.cli.CliClient.callInsertInto(CliClient.java:535)
>   at 
> org.apache.flink.table.client.cli.CliClient.lambda$submitUpdate$0(CliClient.java:231)
>   at java.util.Optional.map(Optional.java:215)
>   at 
> org.apache.flink.table.client.cli.CliClient.submitUpdate(CliClient.java:228)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:129)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:178)
> Caused by: java.lang.RuntimeException: Error while applying rule 
> PushProjectIntoTableSourceScanRule, args 
> [rel#88:FlinkLogicalCalc.LOGICAL(input=RelSubset#87,expr#0..2={inputs},expr#3=IS
>  NOT NULL($t1),user=$t1,rowtime=$t0,$condition=$t3), 
> Scan(table:[default_catalog, default_database, JsonSourceTable], 
> fields:(rowtime, user, event), source:KafkaTableSource(rowtime, user, event))]
>   at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:235)
>   at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:631)
>   at 
> org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:327)
>   at 
> org.apache.flink.table.plan.Optimizer.runVolcanoPlanner(Optimizer.scala:280)
>   at 
> org.apache.flink.table.plan.Optimizer.optimizeLogicalPlan(Optimizer.scala:199)
>   at 
> org.apache.flink.table.plan.StreamOptimizer.optimize(StreamOptimizer.scala:66)
>   at 
> org.apache.flink.table.planner.StreamPlanner.writeToUpsertSink(StreamPlanner.scala:353)
>   at 
> org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$writeToSink(StreamPlanner.scala:281)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:166)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:145)
>   at scala.Option.map(Option.scala:146)
>   at 
> org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$translate(StreamPlanner.scala:145)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117)
>   at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>   at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>   at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
>   at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>   at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
>   at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>   at 
> org.apache.flink.table.planner.StreamPlanner.translate(StreamPlanner.scala:117)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:680)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:493)
>   at 
> org.apache.flink.table.api.java.internal.StreamTableEnvironmentImpl.sqlUpdate(StreamTableEnvironmentImpl.java:331)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$applyUpdate$15(LocalExecutor.java:690)
>   at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:240)
>   at 
> 

[jira] [Closed] (FLINK-15316) SQL Client end-to-end test (Old planner) failed on Travis

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao closed FLINK-15316.

Resolution: Fixed

cc: [~dwysakowicz]

> SQL Client end-to-end test (Old planner) failed on Travis
> -
>
> Key: FLINK-15316
> URL: https://issues.apache.org/jira/browse/FLINK-15316
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client, Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.10.0
>
>
> https://api.travis-ci.org/v3/job/626286501/log.txt
> {noformat}
> Checking for errors...
> Found error in log files:
> ...
> org.apache.flink.table.client.gateway.SqlExecutionException: Invalid SQL 
> update statement.
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.applyUpdate(LocalExecutor.java:697)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdateInternal(LocalExecutor.java:576)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdate(LocalExecutor.java:527)
>   at 
> org.apache.flink.table.client.cli.CliClient.callInsertInto(CliClient.java:535)
>   at 
> org.apache.flink.table.client.cli.CliClient.lambda$submitUpdate$0(CliClient.java:231)
>   at java.util.Optional.map(Optional.java:215)
>   at 
> org.apache.flink.table.client.cli.CliClient.submitUpdate(CliClient.java:228)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:129)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:178)
> Caused by: java.lang.RuntimeException: Error while applying rule 
> PushProjectIntoTableSourceScanRule, args 
> [rel#88:FlinkLogicalCalc.LOGICAL(input=RelSubset#87,expr#0..2={inputs},expr#3=IS
>  NOT NULL($t1),user=$t1,rowtime=$t0,$condition=$t3), 
> Scan(table:[default_catalog, default_database, JsonSourceTable], 
> fields:(rowtime, user, event), source:KafkaTableSource(rowtime, user, event))]
>   at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:235)
>   at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:631)
>   at 
> org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:327)
>   at 
> org.apache.flink.table.plan.Optimizer.runVolcanoPlanner(Optimizer.scala:280)
>   at 
> org.apache.flink.table.plan.Optimizer.optimizeLogicalPlan(Optimizer.scala:199)
>   at 
> org.apache.flink.table.plan.StreamOptimizer.optimize(StreamOptimizer.scala:66)
>   at 
> org.apache.flink.table.planner.StreamPlanner.writeToUpsertSink(StreamPlanner.scala:353)
>   at 
> org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$writeToSink(StreamPlanner.scala:281)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:166)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:145)
>   at scala.Option.map(Option.scala:146)
>   at 
> org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$translate(StreamPlanner.scala:145)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117)
>   at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>   at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>   at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
>   at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>   at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
>   at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>   at 
> org.apache.flink.table.planner.StreamPlanner.translate(StreamPlanner.scala:117)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:680)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:493)
>   at 
> org.apache.flink.table.api.java.internal.StreamTableEnvironmentImpl.sqlUpdate(StreamTableEnvironmentImpl.java:331)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$applyUpdate$15(LocalExecutor.java:690)
>   at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:240)
> 

[jira] [Closed] (FLINK-15316) SQL Client end-to-end test (Old planner) failed on Travis

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao closed FLINK-15316.

Resolution: Duplicate

> SQL Client end-to-end test (Old planner) failed on Travis
> -
>
> Key: FLINK-15316
> URL: https://issues.apache.org/jira/browse/FLINK-15316
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client, Tests
>Affects Versions: 1.10.0
>Reporter: Gary Yao
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.10.0
>
>
> https://api.travis-ci.org/v3/job/626286501/log.txt
> {noformat}
> Checking for errors...
> Found error in log files:
> ...
> org.apache.flink.table.client.gateway.SqlExecutionException: Invalid SQL 
> update statement.
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.applyUpdate(LocalExecutor.java:697)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdateInternal(LocalExecutor.java:576)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.executeUpdate(LocalExecutor.java:527)
>   at 
> org.apache.flink.table.client.cli.CliClient.callInsertInto(CliClient.java:535)
>   at 
> org.apache.flink.table.client.cli.CliClient.lambda$submitUpdate$0(CliClient.java:231)
>   at java.util.Optional.map(Optional.java:215)
>   at 
> org.apache.flink.table.client.cli.CliClient.submitUpdate(CliClient.java:228)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:129)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:104)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:178)
> Caused by: java.lang.RuntimeException: Error while applying rule 
> PushProjectIntoTableSourceScanRule, args 
> [rel#88:FlinkLogicalCalc.LOGICAL(input=RelSubset#87,expr#0..2={inputs},expr#3=IS
>  NOT NULL($t1),user=$t1,rowtime=$t0,$condition=$t3), 
> Scan(table:[default_catalog, default_database, JsonSourceTable], 
> fields:(rowtime, user, event), source:KafkaTableSource(rowtime, user, event))]
>   at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:235)
>   at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:631)
>   at 
> org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:327)
>   at 
> org.apache.flink.table.plan.Optimizer.runVolcanoPlanner(Optimizer.scala:280)
>   at 
> org.apache.flink.table.plan.Optimizer.optimizeLogicalPlan(Optimizer.scala:199)
>   at 
> org.apache.flink.table.plan.StreamOptimizer.optimize(StreamOptimizer.scala:66)
>   at 
> org.apache.flink.table.planner.StreamPlanner.writeToUpsertSink(StreamPlanner.scala:353)
>   at 
> org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$writeToSink(StreamPlanner.scala:281)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:166)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:145)
>   at scala.Option.map(Option.scala:146)
>   at 
> org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$translate(StreamPlanner.scala:145)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117)
>   at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117)
>   at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>   at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>   at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
>   at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>   at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
>   at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>   at 
> org.apache.flink.table.planner.StreamPlanner.translate(StreamPlanner.scala:117)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:680)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:493)
>   at 
> org.apache.flink.table.api.java.internal.StreamTableEnvironmentImpl.sqlUpdate(StreamTableEnvironmentImpl.java:331)
>   at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$applyUpdate$15(LocalExecutor.java:690)
>   at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:240)
>   at 
> 

[jira] [Updated] (FLINK-14007) Add doc for how to using Java user-defined source/sink in Python API

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14007?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-14007:
-
Fix Version/s: (was: 1.11.0)

> Add doc for how to using Java user-defined source/sink in Python API
> 
>
> Key: FLINK-14007
> URL: https://issues.apache.org/jira/browse/FLINK-14007
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python, Documentation
>Reporter: sunjincheng
>Assignee: Hequn Cheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Currently, user can not find out the doc for how to using Java user-defined 
> source/sink in Python API. So we should add the detail doc.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15061) create/alter table/databases properties should be case sensitive stored in catalog

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15061:
-
Fix Version/s: (was: 1.11.0)

> create/alter table/databases properties should be case sensitive stored in 
> catalog
> --
>
> Key: FLINK-15061
> URL: https://issues.apache.org/jira/browse/FLINK-15061
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Terry Wang
>Assignee: Terry Wang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Now in the class `SqlToOperationConverter`, when creating a table the logic 
> will convert all properties key to lower format, which will cause the 
> properties stored in catalog to lose the case style and not intuitively be 
> observed to user.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15318) RocksDBWriteBatchPerformanceTest.benchMark fails on ppc64le

2019-12-18 Thread Siddhesh Ghadi (Jira)
Siddhesh Ghadi created FLINK-15318:
--

 Summary: RocksDBWriteBatchPerformanceTest.benchMark fails on 
ppc64le
 Key: FLINK-15318
 URL: https://issues.apache.org/jira/browse/FLINK-15318
 Project: Flink
  Issue Type: Bug
  Components: Benchmarks, Runtime / State Backends
 Environment: arch: ppc64le
os: rhel7.6, ubuntu 18.04
jdk: 8, 11
mvn: 3.3.9, 3.6.2
Reporter: Siddhesh Ghadi
 Attachments: surefire-report.txt

RocksDBWriteBatchPerformanceTest.benchMark fails due to TestTimedOut, however 
when test-timeout is increased from 2s to 5s in 
org/apache/flink/contrib/streaming/state/benchmark/RocksDBWriteBatchPerformanceTest.java:75,
 it passes. Is this acceptable solution?

Note: Tests are ran inside a container.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-14170) Support hadoop < 2.7 with StreamingFileSink.BulkFormatBuilder

2019-12-18 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas closed FLINK-14170.
--
Resolution: Fixed

Merged on master with f1c4ddf00a3c40dd451723d80e21adeb8b31fa2e
and on release-1.10 with 9c5313e1f634ade92d645dd20ea8b92552a0a3d0

> Support hadoop < 2.7 with StreamingFileSink.BulkFormatBuilder
> -
>
> Key: FLINK-14170
> URL: https://issues.apache.org/jira/browse/FLINK-14170
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem
>Affects Versions: 1.8.0, 1.8.1, 1.8.2, 1.9.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently, StreamingFileSink is supported only with Hadoop >= 2.7 
> irrespective of Row/bulk format builder. This restriction is due to truncate 
> is not supported in  Hadoop < 2.7
> However, BulkFormatBuilder does not use truncate method to restore the file. 
> So the restricting StreamingFileSink.BulkFormatBuilder to be used only with 
> Hadoop >= 2.7 is not necessary.
> So requested improvement is to remove the precondition on 
> HadoopRecoverableWriter and allow  BulkFormatBuilder (Parquet) to be used in 
> Hadoop 2.6 ( Most of the enterprises still on CDH 5.x)
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-14170) Support hadoop < 2.7 with StreamingFileSink.BulkFormatBuilder

2019-12-18 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas reassigned FLINK-14170:
--

Assignee: Bhagavan

> Support hadoop < 2.7 with StreamingFileSink.BulkFormatBuilder
> -
>
> Key: FLINK-14170
> URL: https://issues.apache.org/jira/browse/FLINK-14170
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem
>Affects Versions: 1.8.0, 1.8.1, 1.8.2, 1.9.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently, StreamingFileSink is supported only with Hadoop >= 2.7 
> irrespective of Row/bulk format builder. This restriction is due to truncate 
> is not supported in  Hadoop < 2.7
> However, BulkFormatBuilder does not use truncate method to restore the file. 
> So the restricting StreamingFileSink.BulkFormatBuilder to be used only with 
> Hadoop >= 2.7 is not necessary.
> So requested improvement is to remove the precondition on 
> HadoopRecoverableWriter and allow  BulkFormatBuilder (Parquet) to be used in 
> Hadoop 2.6 ( Most of the enterprises still on CDH 5.x)
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15128) Document support for Hive timestamp type

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15128:
-
Fix Version/s: (was: 1.11.0)

> Document support for Hive timestamp type
> 
>
> Key: FLINK-15128
> URL: https://issues.apache.org/jira/browse/FLINK-15128
> Project: Flink
>  Issue Type: Task
>  Components: Connectors / Hive, Documentation
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14170) Support hadoop < 2.7 with StreamingFileSink.BulkFormatBuilder

2019-12-18 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas updated FLINK-14170:
---
Fix Version/s: 1.10.0

> Support hadoop < 2.7 with StreamingFileSink.BulkFormatBuilder
> -
>
> Key: FLINK-14170
> URL: https://issues.apache.org/jira/browse/FLINK-14170
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem
>Affects Versions: 1.8.0, 1.8.1, 1.8.2, 1.9.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently, StreamingFileSink is supported only with Hadoop >= 2.7 
> irrespective of Row/bulk format builder. This restriction is due to truncate 
> is not supported in  Hadoop < 2.7
> However, BulkFormatBuilder does not use truncate method to restore the file. 
> So the restricting StreamingFileSink.BulkFormatBuilder to be used only with 
> Hadoop >= 2.7 is not necessary.
> So requested improvement is to remove the precondition on 
> HadoopRecoverableWriter and allow  BulkFormatBuilder (Parquet) to be used in 
> Hadoop 2.6 ( Most of the enterprises still on CDH 5.x)
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15179) Kubernetes should not have a CustomCommandLine.

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15179?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15179:
-
Fix Version/s: 1.10.0

> Kubernetes should not have a CustomCommandLine.
> ---
>
> Key: FLINK-15179
> URL: https://issues.apache.org/jira/browse/FLINK-15179
> Project: Flink
>  Issue Type: Sub-task
>  Components: Deployment / Kubernetes
>Affects Versions: 1.10.0
>Reporter: Kostas Kloudas
>Assignee: Kostas Kloudas
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> As part of FLIP-73, all command line options are mapped to config options. 
> Given this 1-to-1 mapping, the Kubernetes command line could simply forward 
> the command line arguments to ConfigOptions directly, instead of introducing 
> new command line options. In this case, the user is expected to simply write:
>  
> {\{bin/run -e (or --executor) kubernetes-session-cluster -D 
> kubernetes.container.image=MY_IMAGE ...}} 
> and the CLI will parse the -e to figure out the correct 
> {{ClusterClientFactory}} and {{ExecutorFactory}} and then forward to that the 
> config options specified with {{-D}}. 
> For this, we need to introduce a {{GenericCustomCommandLine}} that simply 
> forward the specified parameters to the executors.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15190) add documentation for DDL in FLIP-69

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15190?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15190:
-
Fix Version/s: (was: 1.11.0)

> add documentation for DDL in FLIP-69
> 
>
> Key: FLINK-15190
> URL: https://issues.apache.org/jira/browse/FLINK-15190
> Project: Flink
>  Issue Type: Task
>  Components: Documentation
>Reporter: Bowen Li
>Assignee: Terry Wang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> in 
> https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/table/sql.html#ddl



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15203) rephrase Hive's data types doc

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15203?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15203:
-
Fix Version/s: (was: 1.11.0)

> rephrase Hive's data types doc
> --
>
> Key: FLINK-15203
> URL: https://issues.apache.org/jira/browse/FLINK-15203
> Project: Flink
>  Issue Type: Task
>  Components: Documentation
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15228) Drop vendor specific deployment documentation

2019-12-18 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15228?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15228:
-
Fix Version/s: (was: 1.11.0)

> Drop vendor specific deployment documentation
> -
>
> Key: FLINK-15228
> URL: https://issues.apache.org/jira/browse/FLINK-15228
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Seth Wiesman
>Assignee: Seth Wiesman
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Based on a mailing list discussion we want to drop vendor specific deployment 
> documentation
> ml discussion: 
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Drop-vendor-specific-deployment-documentation-td35457.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   >