[GitHub] [flink] flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table
flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table URL: https://github.com/apache/flink/pull/9088#issuecomment-510484364 ## CI report: * eac5f74690ddb0b08cb41b029f5b8ac675e63565 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/118891245) * b22f836f5e8f95a9f376f54c68798eeb14cb1644 : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119054589) * 64007bc344772aa3496f5d8b0a456f73466bbd17 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119417375) * efd6e3189c7ee347335c0ec028603caea2001740 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119441225) * d8b308866574c18cdcd6c89cda7f5837ef50352d : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119955505) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
flinkbot edited a comment on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase URL: https://github.com/apache/flink/pull/9191#issuecomment-513649015 ## CI report: * ddd758fdc5714f75114c1cd35cd61e45091cf736 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119960105) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961 ## CI report: * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119766053) * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119948346) * 7536e2fb5a510a0ca4dfb46476a79307011e9286 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119954049) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot commented on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
flinkbot commented on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase URL: https://github.com/apache/flink/pull/9191#issuecomment-513649015 ## CI report: * ddd758fdc5714f75114c1cd35cd61e45091cf736 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119960105) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot commented on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
flinkbot commented on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase URL: https://github.com/apache/flink/pull/9191#issuecomment-513647351 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Updated] (FLINK-11767) Update TypeSerializerSnapshotMigrationTestBase and subclasses for 1.8
[ https://issues.apache.org/jira/browse/FLINK-11767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-11767: --- Labels: pull-request-available (was: ) > Update TypeSerializerSnapshotMigrationTestBase and subclasses for 1.8 > - > > Key: FLINK-11767 > URL: https://issues.apache.org/jira/browse/FLINK-11767 > Project: Flink > Issue Type: Sub-task > Components: API / Type Serialization System, Tests >Affects Versions: 1.8.0 >Reporter: vinoyang >Assignee: Tzu-Li (Gordon) Tai >Priority: Blocker > Labels: pull-request-available > Fix For: 1.9.0 > > > Update {{TypeSerializerSnapshotMigrationTestBase}} and subclasses to cover > restoring from Flink 1.8. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-10806) Support multiple consuming offsets when discovering a new topic
[ https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889932#comment-16889932 ] Jiayi Liao commented on FLINK-10806: [~becket_qin] Thanks for pointing this out. What you described is more like a migration of state(migrate T1 state into Pipeline B), which reminds me that the savepoint connector that community is building now, can't support the state migration of kafka(maybe you can help build it ^_^). I've already support the simple strategy(earliest/latest) in our own flink branch. But your concerns make sense so I can't give a conclusion now. Maybe we can wait until we get more user cases and scenarios? > Support multiple consuming offsets when discovering a new topic > --- > > Key: FLINK-10806 > URL: https://issues.apache.org/jira/browse/FLINK-10806 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka >Affects Versions: 1.6.2, 1.8.1 >Reporter: Jiayi Liao >Assignee: Jiayi Liao >Priority: Major > > In KafkaConsumerBase, we discover the TopicPartitions and compare them with > the restoredState. It's reasonable when a topic's partitions scaled. However, > if we add a new topic which has too much data and restore the Flink program, > the data of the new topic will be consumed from the start, which may not be > what we want. I think this should be an option for developers. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] tzulitai opened a new pull request #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
tzulitai opened a new pull request #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase URL: https://github.com/apache/flink/pull/9191 #What is the purpose of the change This PR introduces a new `TypeSerializerUprgadeTestBase` to replace the older `TypeSerializerSnapshotMigrationTestBase`. The new test base allows subclasses to define different versions of the same runtime class, by relocating different classes to be under the same name using the `ClassRelocator`. This PR already tries the new test base out with the `PojoSerializer`. ## Verifying this change This is a test related change. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): no - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: no - The serializers: no - The runtime per-record code paths (performance sensitive): no - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: no - The S3 file system connector: no ## Documentation - Does this pull request introduce a new feature? no - If yes, how is the feature documented? n/a This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
flinkbot edited a comment on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#issuecomment-513625214 ## CI report: * 37027083148923256920cbace9ef1f55b0bced5b : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119952271) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Closed] (FLINK-13094) Provide an easy way to read timers using the State Processor API
[ https://issues.apache.org/jira/browse/FLINK-13094?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tzu-Li (Gordon) Tai closed FLINK-13094. --- Resolution: Fixed Fix Version/s: 1.9.0 Merged. master (1.10.0): 752e7ae6a43455234606a9a7da50f40182b2db51 1.9.0: b8eec2b822af8648c73422bcb86ae6b3822420a9 > Provide an easy way to read timers using the State Processor API > > > Key: FLINK-13094 > URL: https://issues.apache.org/jira/browse/FLINK-13094 > Project: Flink > Issue Type: Sub-task > Components: API / DataStream, Runtime / State Backends >Reporter: Tzu-Li (Gordon) Tai >Assignee: Seth Wiesman >Priority: Major > Labels: pull-request-available > Fix For: 1.9.0 > > Time Spent: 20m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13293) Add state processor api to opt/ directory in flink-dist
[ https://issues.apache.org/jira/browse/FLINK-13293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tzu-Li (Gordon) Tai updated FLINK-13293: Affects Version/s: (was: 1.10.0) (was: 1.9.0) > Add state processor api to opt/ directory in flink-dist > --- > > Key: FLINK-13293 > URL: https://issues.apache.org/jira/browse/FLINK-13293 > Project: Flink > Issue Type: Bug > Components: Build System >Reporter: Seth Wiesman >Assignee: Seth Wiesman >Priority: Major > Labels: pull-request-available > Fix For: 1.9.0 > > Time Spent: 20m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Closed] (FLINK-13293) Add state processor api to opt/ directory in flink-dist
[ https://issues.apache.org/jira/browse/FLINK-13293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Tzu-Li (Gordon) Tai closed FLINK-13293. --- Resolution: Fixed Fix Version/s: (was: 1.10.0) Merged. master (1.10.0): c8af3c95dbac76458c01bd50e4d6f85aeca3e89a 1.9.0: 752e7ae6a43455234606a9a7da50f40182b2db51 > Add state processor api to opt/ directory in flink-dist > --- > > Key: FLINK-13293 > URL: https://issues.apache.org/jira/browse/FLINK-13293 > Project: Flink > Issue Type: Bug > Components: Build System >Affects Versions: 1.9.0, 1.10.0 >Reporter: Seth Wiesman >Assignee: Seth Wiesman >Priority: Major > Labels: pull-request-available > Fix For: 1.9.0 > > Time Spent: 20m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …
flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in … URL: https://github.com/apache/flink/pull/9165#issuecomment-512797712 ## CI report: * c51d878e9c6c9ccce6b1d67969aa029395acc4ea : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119628900) * a3ad9538892a23a689945f6cbf315625364dc756 : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119951986) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table
flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table URL: https://github.com/apache/flink/pull/9088#issuecomment-510484364 ## CI report: * eac5f74690ddb0b08cb41b029f5b8ac675e63565 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/118891245) * b22f836f5e8f95a9f376f54c68798eeb14cb1644 : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119054589) * 64007bc344772aa3496f5d8b0a456f73466bbd17 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119417375) * efd6e3189c7ee347335c0ec028603caea2001740 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119441225) * d8b308866574c18cdcd6c89cda7f5837ef50352d : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119955505) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#issuecomment-510770597 ## CI report: * 6afb0645f386b618ddc1e9d86add2a22967676d6 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/118887349) * e5f48986b0ceb6bc774f33f3086898a5e6afb06c : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119123700) * a75172de1ff1d6c2a4b12dde013258fd7a36fc00 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119951727) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serialize
flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serializer restore in BaseArray/Map serializer and add tests for complex data formats URL: https://github.com/apache/flink/pull/9139#issuecomment-512166499 ## CI report: * 22609f1b6271176affec70926b89b4730451568a : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119436577) * 7dbec40e5321c5ffcaae89c8716851a8a1248519 : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119572205) * 85740dd1f663191b172b2601c2e4a0a4c5167ad1 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119625173) * eeb0b5cea4c0a633e76ef61667bdb52d6bad520f : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119636253) * 3519a3632830d4edc2b90a1cc2b0e467ebcb823c : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119766271) * 8e1e152e94f0114237baf148e8e9052afb9c2756 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119953465) * a2a42395a583f1268de18486ee0beef6eeb491cc : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119956594) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] asfgit closed pull request #9094: [FLINK-13094][state-processor-api] Provide an easy way to read timers using the State Processor API
asfgit closed pull request #9094: [FLINK-13094][state-processor-api] Provide an easy way to read timers using the State Processor API URL: https://github.com/apache/flink/pull/9094 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections
[ https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889905#comment-16889905 ] TisonKun commented on FLINK-10052: -- Yes. LeaderContender(jm, rm, dispatcher, etc.) in FLINK is all subclass of {{FencedRpcEndpoint}}, which means that rpc messages of different leadership epoch are fenced. > Tolerate temporarily suspended ZooKeeper connections > > > Key: FLINK-10052 > URL: https://issues.apache.org/jira/browse/FLINK-10052 > Project: Flink > Issue Type: Improvement > Components: Runtime / Coordination >Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1 >Reporter: Till Rohrmann >Assignee: Dominik Wosiński >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > This issue results from FLINK-10011 which uncovered a problem with Flink's HA > recovery and proposed the following solution to harden Flink: > The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator > recipe for leader election. The leader latch revokes leadership in case of a > suspended ZooKeeper connection. This can be premature in case that the system > can reconnect to ZooKeeper before its session expires. The effect of the lost > leadership is that all jobs will be canceled and directly restarted after > regaining the leadership. > Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper > connection, it would be better to wait until the ZooKeeper connection is > LOST. That way we would allow the system to reconnect and not lose the > leadership. This could be achievable by using Curator's {{LeaderSelector}} > instead of the {{LeaderLatch}}. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] asfgit closed pull request #9133: [FLINK-13293][state-processor-api][build] Add state processor api to opt/ directory in flink-dist
asfgit closed pull request #9133: [FLINK-13293][state-processor-api][build] Add state processor api to opt/ directory in flink-dist URL: https://github.com/apache/flink/pull/9133 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] tzulitai commented on issue #9094: [FLINK-13094][state-processor-api] Provide an easy way to read timers using the State Processor API
tzulitai commented on issue #9094: [FLINK-13094][state-processor-api] Provide an easy way to read timers using the State Processor API URL: https://github.com/apache/flink/pull/9094#issuecomment-513636515 Merging ... This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] tzulitai commented on issue #9133: [FLINK-13293][state-processor-api][build] Add state processor api to opt/ directory in flink-dist
tzulitai commented on issue #9133: [FLINK-13293][state-processor-api][build] Add state processor api to opt/ directory in flink-dist URL: https://github.com/apache/flink/pull/9133#issuecomment-513636024 +1, merging ... This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Closed] (FLINK-13346) indexdb reporter parse data type error
[ https://issues.apache.org/jira/browse/FLINK-13346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yun Tang closed FLINK-13346. Resolution: Duplicate > indexdb reporter parse data type error > -- > > Key: FLINK-13346 > URL: https://issues.apache.org/jira/browse/FLINK-13346 > Project: Flink > Issue Type: Bug >Affects Versions: 1.7.2, 1.8.0, 1.8.1 > Environment: Yarn cluster >Reporter: Jake.zhang >Priority: Minor > > unable to parse > 'taskmanager_job_task_operator_KafkaConsumer_sync-time-max,host=dn45,job_id=f9bc38f9d31ac689715c055a58905288,job_name=network_monitor_save_phoenix,operator_id=cbc357ccb763df2852fee8c4fc7d55f2,operator_name=Source:\ > Custom\ > Source,subtask_index=0,task_attempt_id=3462928b2995a24303eefad4ebfb9f4f,task_attempt_num=0,task_id=cbc357ccb763df2852fee8c4fc7d55f2,task_name=Source:\ > Custom\ Source\ ->\ Map\ ->\ Filter\ ->\ Sink:\ > Unnamed,tm_id=container_e65_1560932557384_276950_01_000647 value=-∞ > 156376155872700': invalid number dropped=0 > at > org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionFromErrorMessage(InfluxDBException.java:147) > at > org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionForErrorState(InfluxDBException.java:173) > at > org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.execute(InfluxDBImpl.java:796) > at > org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.write(InfluxDBImpl.java:455) > at > org.apache.flink.metrics.influxdb.InfluxdbReporter.report(InfluxdbReporter.java:97) > at > org.apache.flink.runtime.metrics.MetricRegistryImpl$ReporterTask.run(MetricRegistryImpl.java:427) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections
[ https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889901#comment-16889901 ] Quan Shi commented on FLINK-10052: -- hi [~Tison] Sounds reasonable. Elect a new leader and recovery from checkpoints is a concise way for HA. What happens to FLINK if zookeeper split-brain or zookeeper cluster shutdown? Does FLINK has fencing mechanism like HDFS? > Tolerate temporarily suspended ZooKeeper connections > > > Key: FLINK-10052 > URL: https://issues.apache.org/jira/browse/FLINK-10052 > Project: Flink > Issue Type: Improvement > Components: Runtime / Coordination >Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1 >Reporter: Till Rohrmann >Assignee: Dominik Wosiński >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > This issue results from FLINK-10011 which uncovered a problem with Flink's HA > recovery and proposed the following solution to harden Flink: > The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator > recipe for leader election. The leader latch revokes leadership in case of a > suspended ZooKeeper connection. This can be premature in case that the system > can reconnect to ZooKeeper before its session expires. The effect of the lost > leadership is that all jobs will be canceled and directly restarted after > regaining the leadership. > Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper > connection, it would be better to wait until the ZooKeeper connection is > LOST. That way we would allow the system to reconnect and not lose the > leadership. This could be achievable by using Curator's {{LeaderSelector}} > instead of the {{LeaderLatch}}. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] lirui-apache commented on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table
lirui-apache commented on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table URL: https://github.com/apache/flink/pull/9088#issuecomment-513633987 Changed to test with table env. This PR needs to wait until we have #9181 merged. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serialize
flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serializer restore in BaseArray/Map serializer and add tests for complex data formats URL: https://github.com/apache/flink/pull/9139#issuecomment-512166499 ## CI report: * 22609f1b6271176affec70926b89b4730451568a : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119436577) * 7dbec40e5321c5ffcaae89c8716851a8a1248519 : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119572205) * 85740dd1f663191b172b2601c2e4a0a4c5167ad1 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119625173) * eeb0b5cea4c0a633e76ef61667bdb52d6bad520f : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119636253) * 3519a3632830d4edc2b90a1cc2b0e467ebcb823c : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119766271) * 8e1e152e94f0114237baf148e8e9052afb9c2756 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119953465) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (FLINK-13344) Translate "How to Contribute" page into Chinese.
[ https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889895#comment-16889895 ] Jeff Yang commented on FLINK-13344: --- I got it . Thanks [~jark] > Translate "How to Contribute" page into Chinese. > - > > Key: FLINK-13344 > URL: https://issues.apache.org/jira/browse/FLINK-13344 > Project: Flink > Issue Type: Sub-task > Components: chinese-translation, Project Website >Reporter: Jark Wu >Assignee: WangHengWei >Priority: Major > > The page is https://flink.apache.org/zh/contributing/how-to-contribute.html > The markdown file is located in > https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md > Before start working on this, please read translation guideline: > https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-13344) Translate "How to Contribute" page into Chinese.
[ https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889894#comment-16889894 ] WangHengWei commented on FLINK-13344: - [~jark], thanks very much. I'm gonna do it.:D > Translate "How to Contribute" page into Chinese. > - > > Key: FLINK-13344 > URL: https://issues.apache.org/jira/browse/FLINK-13344 > Project: Flink > Issue Type: Sub-task > Components: chinese-translation, Project Website >Reporter: Jark Wu >Assignee: WangHengWei >Priority: Major > > The page is https://flink.apache.org/zh/contributing/how-to-contribute.html > The markdown file is located in > https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md > Before start working on this, please read translation guideline: > https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections
[ https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889893#comment-16889893 ] TisonKun commented on FLINK-10052: -- [~quan] Supposed you have jm-1 and jm-2 execute the same job but run on different cluster. jm-1 is the previous leader and thus jm-2 is inactive. Now jm-1 lost connection with zookeeper and jm-2 became the leader, recovering from job graph store and checkpoint store, started executing the job. If we didn't shutdown jm-1 and supposed it would be shutdown due to zk communication failure, not only there were two jms running for the same job thus the same job was executed twice, but also they might conflict writing states. To sum up, if a jm lost connection with zk, there is possibly another jm running for the same job. We don't want to see 2 jms running for the same job at the same time, right? If you're worry about the performance, when you executing a streaming job, flink can recover it from checkpoints and catch up quickly. > Tolerate temporarily suspended ZooKeeper connections > > > Key: FLINK-10052 > URL: https://issues.apache.org/jira/browse/FLINK-10052 > Project: Flink > Issue Type: Improvement > Components: Runtime / Coordination >Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1 >Reporter: Till Rohrmann >Assignee: Dominik Wosiński >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > This issue results from FLINK-10011 which uncovered a problem with Flink's HA > recovery and proposed the following solution to harden Flink: > The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator > recipe for leader election. The leader latch revokes leadership in case of a > suspended ZooKeeper connection. This can be premature in case that the system > can reconnect to ZooKeeper before its session expires. The effect of the lost > leadership is that all jobs will be canceled and directly restarted after > regaining the leadership. > Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper > connection, it would be better to wait until the ZooKeeper connection is > LOST. That way we would allow the system to reconnect and not lose the > leadership. This could be achievable by using Curator's {{LeaderSelector}} > instead of the {{LeaderLatch}}. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961 ## CI report: * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119766053) * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119948346) * 7536e2fb5a510a0ca4dfb46476a79307011e9286 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119954049) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] KurtYoung commented on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
KurtYoung commented on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#issuecomment-51363 LGTM, waiting travis pass This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUE
KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#discussion_r305665394 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java ## @@ -455,6 +455,15 @@ protected void restartTasks(final Set verticesToRestart) { this.delegate = checkNotNull(delegate); } + @Override + public CompletableFuture allocateBatchSlot( Review comment: Thanks for the clarification! This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "
[ https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jingsong Lee updated FLINK-13289: - Priority: Blocker (was: Major) > Blink Planner JDBCUpsertTableSink : UnsupportedOperationException > "JDBCUpsertTableSink can not support " > > > Key: FLINK-13289 > URL: https://issues.apache.org/jira/browse/FLINK-13289 > Project: Flink > Issue Type: Sub-task > Components: Connectors / JDBC, Table SQL / Planner >Affects Versions: 1.9.0 >Reporter: LakeShen >Priority: Blocker > Fix For: 1.9.0, 1.10.0 > > > Hi , in flink-jdbc connector module, I change the Flink planner to Blink > planner to test all test case,because we want to use Blank planner in our > program. When I test the JDBCUpsertTableSinkITCase class , the method > testUpsert throw the exception: > {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can > not support {color} > I saw the src code,in Flink planner , the StreamPlanner set the > JDBCUpsertTableSink' keyFields, > but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' > keyFields,so JDBCUpsertTableSink keyFields is null, when execute > JDBCUpsertTableSink newFormat(), > it thrown the exception. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13342) Flink1.9 document number error
[ https://issues.apache.org/jira/browse/FLINK-13342?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13342: Priority: Critical (was: Major) > Flink1.9 document number error > -- > > Key: FLINK-13342 > URL: https://issues.apache.org/jira/browse/FLINK-13342 > Project: Flink > Issue Type: Bug >Affects Versions: 1.9.0 >Reporter: hanfei >Priority: Critical > Fix For: 1.9.0 > > > link:https://flink.apache.org/ > description:Document menu flink1.9(Snapshot) jump to flink1.10(Snapshot) > bug:Flink1.9 document number error -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Comment Edited] (FLINK-13242) StandaloneResourceManagerTest fails on travis
[ https://issues.apache.org/jira/browse/FLINK-13242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889630#comment-16889630 ] Xintong Song edited comment on FLINK-13242 at 7/22/19 3:45 AM: --- Hi [~azagrebin], I think I found the problem. In _StandaloneResourceManager#initialize()_ it uses _getMainThreadExecutor()_ to execute _setFailUnfulfillableRequest()_. However, before _setFailUnfulfillableRequest()_ is executed, the main thread executor of the resource manager might be replaced by a new one when it accepts the granted leader ship, leading to _setFailUnfulfillableRequest()_ never being executed. This only happens when the _StandaloneResourceManager#initialize()_ is invoked before _TestingLeaderElectionService#isLeader()_. The problem can be re-produced and verified as follows: * Add logs in _StandaloneResourceManager#initialize()_, _TestingLeaderElectionService#isLeader()_ and _ResourceManager#setFailUnfulfillableRequest()_, and run the test. In most cases, you should see _TestingLeaderElectionService#isLeader()_ invoked before _StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is invoked, and the test should pass. * Add a short sleep time (in my case 100ms) in _MockResourceManagerRuntimeServices#grantLeadership()_ before _rmLeaderElectionService.isLeader()_, and run the test again. Now you should see _TestingLeaderElectionService#isLeader()_ invoked after _StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is never invoked, and the test should fail. * Add another short sleep time (also 100ms in my case) in _StandaloneResourceManager#initialize()_, inside the _getRpcService().getScheduledExecutor().schedule()_ block, right before _getMainThreadExecutor()_. This should change the order of invoking back and fix the failure. * If you invoke _getMainThreadExecutor()_ twice in _StandaloneResourceManager#initialize()_, once before the sleep and the other after it, and print out the fetched main thread executors, you should find that they are two different objects. * Now if you remove the sleep in _StandaloneResourceManager#initialize()_, you should see the printed two main thread executors are the same object, and the test is broken again. I'm thinking that maybe _setFailUnfulfillableRequest(true)_ does not need to be invoked on the PRC main thread. Instead of calling on the main thread executor, I tried call _setFailUnfulfillableRequest(true)_ directly in the _getRpcService().getScheduledExecutor().schedule()_ block in _StandaloneResourceManager#initialize()_ and it fixes the problem. I think we do not care whether the _setFailUnfulfillableRequest(true)_ happens on main thread or not in production, as long as it eventually get invoked. And for this test case, we may have a bit inconsistency that after _setFailUnfulfillableRequest(true)_ the _isFailingUnfulfillableRequest()_ may not get the correct result immediately, which I think is acceptable and the 10s timeout for _assertHappensUntil()_ should be long enough to catch the invoking of _setFailUnfulfillableRequest(true)_ eventually. What do you think? was (Author: xintongsong): Hi [~azagrebin], I think I found the problem. In _StandaloneResourceManager#initialize()_ it uses _getMainThreadExecutor()_ to execute _setFailUnfulfillableRequest()_. However, before _setFailUnfulfillableRequest()_ is executed, the main thread executor of the resource manager might be replaced by a new one when it accepts the granted leader ship, leading to _setFailUnfulfillableRequest()_ never being executed. This only happens when the _StandaloneResourceManager#initialize()_ is invoked before _TestingLeaderElectionService#isLeader()_. The problem can be re-produced and verified as follows: * Add logs in _StandaloneResourceManager#initialize()_, __ _TestingLeaderElectionService#isLeader()_ and _ResourceManager#setFailUnfulfillableRequest()_, and run the test. In most cases, you should see _TestingLeaderElectionService#isLeader()_ invoked before __ _StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is invoked, and the test should pass. * Add a short sleep time (in my case 100ms) in _MockResourceManagerRuntimeServices#grantLeadership()_ before _rmLeaderElectionService.isLeader()_, and run the test again. Now you should see _TestingLeaderElectionService#isLeader()_ invoked after __ _StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is never invoked, and the test should fail. * Add another short sleep time (also 100ms in my case) in _StandaloneResourceManager#initialize()_, inside __ the _getRpcService().getScheduledExecutor().schedule()_ block, right before _getMainThreadExecutor()_. This should change the order of invoking back and fix the failure. * If you invoke _getMainThreadExecutor()_ twice in
[jira] [Commented] (FLINK-13318) Blink planner tests failing on Scala 2.12
[ https://issues.apache.org/jira/browse/FLINK-13318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889891#comment-16889891 ] Jark Wu commented on FLINK-13318: - +1 for this approach! > Blink planner tests failing on Scala 2.12 > - > > Key: FLINK-13318 > URL: https://issues.apache.org/jira/browse/FLINK-13318 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner, Tests >Affects Versions: 1.9.0 >Reporter: Chesnay Schepler >Priority: Blocker > Fix For: 1.9.0 > > > [https://travis-ci.org/apache/flink/builds/559909681] > {code:java} > 13:30:03.531 [INFO] Results: > 13:30:03.531 [INFO] > 13:30:03.533 [ERROR] Failures: > 13:30:03.534 [ERROR] CalcTest.testScalarFunctionAccess:64 planBefore > expected:<...t$giveMeCaseClass$$f[e1bff2b06d8e0e495536102224cfe83().my], > _c1=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$fe1bff2b06d8e0e495536102224cfe83().clazz], > > _c2=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$fe1bff2b06d8e0e495536102224cfe83().my], > > _c3=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$fe1bff2b06d8e0e495536102224cfe83]().clazz]) > +- Logica...> but > was:<...t$giveMeCaseClass$$f[4a420732fc04b1351889eb0e88eb891().my], > _c1=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$f4a420732fc04b1351889eb0e88eb891().clazz], > > _c2=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$f4a420732fc04b1351889eb0e88eb891().my], > > _c3=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$f4a420732fc04b1351889eb0e88eb891]().clazz]) > +- Logica...> > 13:30:03.534 [ERROR] CalcTest.testSelectFromGroupedTableWithFunctionKey:154 > planBefore > expected:<...alcTest$MyHashCode$$[d14b486109d9dd062ae7c60e0497797]5($2)]) > +- Log...> but > was:<...alcTest$MyHashCode$$[3cd929923219fc59162b13a4941ead4]5($2)]) > +- Log...> > 13:30:03.534 [ERROR] CalcTest.testSelectFunction:109 planBefore > expected:<...alcTest$MyHashCode$$[d14b486109d9dd062ae7c60e0497797]5($2)], > b=[$1]) > +- L...> but > was:<...alcTest$MyHashCode$$[3cd929923219fc59162b13a4941ead4]5($2)], b=[$1]) > +- L...> > 13:30:03.534 [ERROR] CorrelateTest.testCrossJoin:41 planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], > rowType=[Rec...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], > rowType=[Rec...> > 13:30:03.534 [ERROR] CorrelateTest.testCrossJoin2:52 planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2, > _UTF-16LE'$')]...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2, > _UTF-16LE'$')]...> > 13:30:03.534 [ERROR] CorrelateTest.testLeftOuterJoinWithLiteralTrue:74 > planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], > rowType=[Rec...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], > rowType=[Rec...> > 13:30:03.534 [ERROR] > CorrelateTest.testLeftOuterJoinWithoutJoinPredicates:63 planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], > rowType=[Rec...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], > rowType=[Rec...> > 13:30:03.535 [ERROR] JoinTest.testFilterJoinRule:143 planBefore > expected:<...le$JoinTest$Merger$$[223b7380fec29c4077a893c60165d845($2, > org$apache$flink$table$plan$batch$table$JoinTest$Merger$$223b7380fec29c4077a893c60165d845]($2, > $5))]) >+- Lo...> but > was:<...le$JoinTest$Merger$$[d18a3011491fab359eccb50f2d0d9a18($2, > org$apache$flink$table$plan$batch$table$JoinTest$Merger$$d18a3011491fab359eccb50f2d0d9a18]($2, > $5))]) >+- Lo...> > 13:30:03.535 [ERROR] CorrelateStringExpressionTest.testCorrelateJoins1:39 > planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], > rowType=[Rec...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], > rowType=[Rec...> > 13:30:03.535 [ERROR] CorrelateStringExpressionTest.testCorrelateJoins2:45 > planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], > rowType=[Rec...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], > rowType=[Rec...> > 13:30:03.535 [ERROR] CorrelateStringExpressionTest.testCorrelateJoins3:51 > planBefore > expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2, > _UTF-16LE'$')]...> but > was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2, > _UTF-16LE'$')]...> > 13:30:03.535 [ERROR] CorrelateStringExpressionTest.testCorrelateJoins4:57 > planBefore > expected:<...ble$util$TableFunc2$[b3b1f988779be024ed9386bce5019112]($2)], > rowType=[Reco...> but >
[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "
[ https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13289: Fix Version/s: 1.10.0 > Blink Planner JDBCUpsertTableSink : UnsupportedOperationException > "JDBCUpsertTableSink can not support " > > > Key: FLINK-13289 > URL: https://issues.apache.org/jira/browse/FLINK-13289 > Project: Flink > Issue Type: Sub-task > Components: Connectors / JDBC, Table SQL / Planner >Affects Versions: 1.9.0 >Reporter: LakeShen >Priority: Major > Fix For: 1.9.0, 1.10.0 > > > Hi , in flink-jdbc connector module, I change the Flink planner to Blink > planner to test all test case,because we want to use Blank planner in our > program. When I test the JDBCUpsertTableSinkITCase class , the method > testUpsert throw the exception: > {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can > not support {color} > I saw the src code,in Flink planner , the StreamPlanner set the > JDBCUpsertTableSink' keyFields, > but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' > keyFields,so JDBCUpsertTableSink keyFields is null, when execute > JDBCUpsertTableSink newFormat(), > it thrown the exception. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "
[ https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13289: Fix Version/s: 1.9.0 > Blink Planner JDBCUpsertTableSink : UnsupportedOperationException > "JDBCUpsertTableSink can not support " > > > Key: FLINK-13289 > URL: https://issues.apache.org/jira/browse/FLINK-13289 > Project: Flink > Issue Type: Sub-task > Components: Connectors / JDBC, Table SQL / Planner >Affects Versions: 1.9.0 >Reporter: LakeShen >Priority: Major > Fix For: 1.9.0 > > > Hi , in flink-jdbc connector module, I change the Flink planner to Blink > planner to test all test case,because we want to use Blank planner in our > program. When I test the JDBCUpsertTableSinkITCase class , the method > testUpsert throw the exception: > {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can > not support {color} > I saw the src code,in Flink planner , the StreamPlanner set the > JDBCUpsertTableSink' keyFields, > but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' > keyFields,so JDBCUpsertTableSink keyFields is null, when execute > JDBCUpsertTableSink newFormat(), > it thrown the exception. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "
[ https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13289: Component/s: Connectors / JDBC > Blink Planner JDBCUpsertTableSink : UnsupportedOperationException > "JDBCUpsertTableSink can not support " > > > Key: FLINK-13289 > URL: https://issues.apache.org/jira/browse/FLINK-13289 > Project: Flink > Issue Type: Sub-task > Components: Connectors / JDBC, Table SQL / Planner >Affects Versions: 1.9.0 >Reporter: LakeShen >Priority: Major > > Hi , in flink-jdbc connector module, I change the Flink planner to Blink > planner to test all test case,because we want to use Blank planner in our > program. When I test the JDBCUpsertTableSinkITCase class , the method > testUpsert throw the exception: > {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can > not support {color} > I saw the src code,in Flink planner , the StreamPlanner set the > JDBCUpsertTableSink' keyFields, > but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' > keyFields,so JDBCUpsertTableSink keyFields is null, when execute > JDBCUpsertTableSink newFormat(), > it thrown the exception. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "
[ https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889890#comment-16889890 ] Jark Wu commented on FLINK-13289: - Thanks [~kakachen] for digging into this. This is definitely a bug in blink planner. We will fix this in planner. > Blink Planner JDBCUpsertTableSink : UnsupportedOperationException > "JDBCUpsertTableSink can not support " > > > Key: FLINK-13289 > URL: https://issues.apache.org/jira/browse/FLINK-13289 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Affects Versions: 1.9.0 >Reporter: LakeShen >Priority: Major > > Hi , in flink-jdbc connector module, I change the Flink planner to Blink > planner to test all test case,because we want to use Blank planner in our > program. When I test the JDBCUpsertTableSinkITCase class , the method > testUpsert throw the exception: > {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can > not support {color} > I saw the src code,in Flink planner , the StreamPlanner set the > JDBCUpsertTableSink' keyFields, > but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' > keyFields,so JDBCUpsertTableSink keyFields is null, when execute > JDBCUpsertTableSink newFormat(), > it thrown the exception. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serialize
flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serializer restore in BaseArray/Map serializer and add tests for complex data formats URL: https://github.com/apache/flink/pull/9139#issuecomment-512166499 ## CI report: * 22609f1b6271176affec70926b89b4730451568a : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119436577) * 7dbec40e5321c5ffcaae89c8716851a8a1248519 : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119572205) * 85740dd1f663191b172b2601c2e4a0a4c5167ad1 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119625173) * eeb0b5cea4c0a633e76ef61667bdb52d6bad520f : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119636253) * 3519a3632830d4edc2b90a1cc2b0e467ebcb823c : SUCCESS [Build](https://travis-ci.com/flink-ci/flink/builds/119766271) * 8e1e152e94f0114237baf148e8e9052afb9c2756 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119953465) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] godfreyhe commented on a change in pull request #9056: [FLINK-13185] [sql-parser][table-planner] Bump Calcite dependency to 1.20.0 in sql parser & flink planner
godfreyhe commented on a change in pull request #9056: [FLINK-13185] [sql-parser][table-planner] Bump Calcite dependency to 1.20.0 in sql parser & flink planner URL: https://github.com/apache/flink/pull/9056#discussion_r305664376 ## File path: flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/plan/nodes/CommonCorrelate.scala ## @@ -46,7 +47,7 @@ trait CommonCorrelate { inputSchema: RowSchema, udtfTypeInfo: TypeInformation[Any], returnSchema: RowSchema, -joinType: SemiJoinType, +joinType: JoinRelType, Review comment: sorry for late reply, I create a JIRA to fix this: https://issues.apache.org/jira/browse/FLINK-13347 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Created] (FLINK-13347) should handle new JoinRelType(SEMI/ANTI) in switch case
godfrey he created FLINK-13347: -- Summary: should handle new JoinRelType(SEMI/ANTI) in switch case Key: FLINK-13347 URL: https://issues.apache.org/jira/browse/FLINK-13347 Project: Flink Issue Type: Bug Components: Table SQL / Legacy Planner, Table SQL / Planner Reporter: godfrey he Fix For: 1.9.0, 1.10.0 Calcite 1.20 introduces {{SEMI}} & {{ANTI}} to {{JoinRelType}}, blink planner & flink planner should handle them in each switch case -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-13344) Translate "How to Contribute" page into Chinese.
[ https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889888#comment-16889888 ] Jark Wu commented on FLINK-13344: - Hi [~highfei2...@126.com], this is not a duplicate issue with FLINK-13343, this is another page :) Hi [~WangHW], I assigned this issue to you. > Translate "How to Contribute" page into Chinese. > - > > Key: FLINK-13344 > URL: https://issues.apache.org/jira/browse/FLINK-13344 > Project: Flink > Issue Type: Sub-task > Components: chinese-translation, Project Website >Reporter: Jark Wu >Assignee: WangHengWei >Priority: Major > > The page is https://flink.apache.org/zh/contributing/how-to-contribute.html > The markdown file is located in > https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md > Before start working on this, please read translation guideline: > https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Assigned] (FLINK-13344) Translate "How to Contribute" page into Chinese.
[ https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu reassigned FLINK-13344: --- Assignee: WangHengWei > Translate "How to Contribute" page into Chinese. > - > > Key: FLINK-13344 > URL: https://issues.apache.org/jira/browse/FLINK-13344 > Project: Flink > Issue Type: Sub-task > Components: chinese-translation, Project Website >Reporter: Jark Wu >Assignee: WangHengWei >Priority: Major > > The page is https://flink.apache.org/zh/contributing/how-to-contribute.html > The markdown file is located in > https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md > Before start working on this, please read translation guideline: > https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections
[ https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889887#comment-16889887 ] Quan Shi commented on FLINK-10052: -- Hi, all. What if the Job Manager works fine except losing connection with zookeeper? The stream job can continue running. Do we really need to select another active Job Manager and restart the whole job? > Tolerate temporarily suspended ZooKeeper connections > > > Key: FLINK-10052 > URL: https://issues.apache.org/jira/browse/FLINK-10052 > Project: Flink > Issue Type: Improvement > Components: Runtime / Coordination >Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1 >Reporter: Till Rohrmann >Assignee: Dominik Wosiński >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > This issue results from FLINK-10011 which uncovered a problem with Flink's HA > recovery and proposed the following solution to harden Flink: > The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator > recipe for leader election. The leader latch revokes leadership in case of a > suspended ZooKeeper connection. This can be premature in case that the system > can reconnect to ZooKeeper before its session expires. The effect of the lost > leadership is that all jobs will be canceled and directly restarted after > regaining the leadership. > Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper > connection, it would be better to wait until the ZooKeeper connection is > LOST. That way we would allow the system to reconnect and not lose the > leadership. This could be achievable by using Curator's {{LeaderSelector}} > instead of the {{LeaderLatch}}. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] liyafan82 commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUE
liyafan82 commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#discussion_r305663679 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java ## @@ -455,6 +455,15 @@ protected void restartTasks(final Set verticesToRestart) { this.delegate = checkNotNull(delegate); } + @Override + public CompletableFuture allocateBatchSlot( Review comment: Yes. Without this method override, the default implementation of the super class will be called, which is based on SlotProvider#allocateBatchSlot. The default implementation throws an UnsupportedOperationException, which will cause some UT to fail. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUE
KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#discussion_r305663260 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java ## @@ -455,6 +455,15 @@ protected void restartTasks(final Set verticesToRestart) { this.delegate = checkNotNull(delegate); } + @Override + public CompletableFuture allocateBatchSlot( Review comment: Is this necessary? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Created] (FLINK-13346) indexdb reporter parse data type error
Jake.zhang created FLINK-13346: -- Summary: indexdb reporter parse data type error Key: FLINK-13346 URL: https://issues.apache.org/jira/browse/FLINK-13346 Project: Flink Issue Type: Bug Affects Versions: 1.8.1, 1.8.0, 1.7.2 Environment: Yarn cluster Reporter: Jake.zhang unable to parse 'taskmanager_job_task_operator_KafkaConsumer_sync-time-max,host=dn45,job_id=f9bc38f9d31ac689715c055a58905288,job_name=network_monitor_save_phoenix,operator_id=cbc357ccb763df2852fee8c4fc7d55f2,operator_name=Source:\ Custom\ Source,subtask_index=0,task_attempt_id=3462928b2995a24303eefad4ebfb9f4f,task_attempt_num=0,task_id=cbc357ccb763df2852fee8c4fc7d55f2,task_name=Source:\ Custom\ Source\ ->\ Map\ ->\ Filter\ ->\ Sink:\ Unnamed,tm_id=container_e65_1560932557384_276950_01_000647 value=-∞ 156376155872700': invalid number dropped=0 at org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionFromErrorMessage(InfluxDBException.java:147) at org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionForErrorState(InfluxDBException.java:173) at org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.execute(InfluxDBImpl.java:796) at org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.write(InfluxDBImpl.java:455) at org.apache.flink.metrics.influxdb.InfluxdbReporter.report(InfluxdbReporter.java:97) at org.apache.flink.runtime.metrics.MetricRegistryImpl$ReporterTask.run(MetricRegistryImpl.java:427) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] JingsongLi commented on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases
JingsongLi commented on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases URL: https://github.com/apache/flink/pull/9180#issuecomment-513625785 @wuchong @zentol Thanks for your attention. About leak resources, there is a clean thread to invoke `MiniClusterWithClientResource.after` to clean resources. Compared with the past, it only delayed cleaning up resources by 15 seconds. >I have doubts whether this will improve the test times in a significant way, since we're only skipping the creation of ~60 mini clusters. Yeah, we should gauge how valuable this change is. 1.I test on my macbook pro, blink-planner will use 17min before, and use 11min now for `mvn clean verify`. If we just run `mvn verify`, the gap will be more pronounced. 2.But the more accurate test time should be travis triggering, I will update later. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (FLINK-10806) Support multiple consuming offsets when discovering a new topic
[ https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889886#comment-16889886 ] Jiangjie Qin commented on FLINK-10806: -- [~wind_ljy] Thanks for the context. Yes, I agree that the use case you mentioned is a pretty reasonable use case. And the fact currently the newly discovered partitions (they may or may not belong to a new topic) are set to {{EARLIEST_OFFSET}} is a bad behavior. What I am wondering is whether you only need to set the consumer offset to the latest for newly discovered partitions, or you may need to set the offset to some specific offset. This may impact the delivery guarantee. For example, consider the following event sequence: # Initial state: Pipeline A is processing Topic T1. Pipeline B is processing Topic T2. For simplicity, let's assume topic T1 and T2 both only have 1 partition (P0). # Stop pipeline A and pipeline B. Reassign both topic 1 and topic 2 to Pipeline B. # The upstream producers produced a few more messages into T1. # Restart pipeline B to consume from both T1 and T2. In the above case, after restart pipeline B, for T2 the consumption will start from the last committed offset, which is probably what we expect. However, for T1, if the consumption starts from the latest, the messages produced in step 3 will be lost. Is that behavior expected? Do you want to set the starting offset to where pipeline A left in this case? It is quite easy to just always allow users to configure the consumer to either start from earliest or latest for the new partitions if the group offset is missing. But allowing a specific starting offset for a newly discovered topic needs a few more thoughts. > Support multiple consuming offsets when discovering a new topic > --- > > Key: FLINK-10806 > URL: https://issues.apache.org/jira/browse/FLINK-10806 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka >Affects Versions: 1.6.2, 1.8.1 >Reporter: Jiayi Liao >Assignee: Jiayi Liao >Priority: Major > > In KafkaConsumerBase, we discover the TopicPartitions and compare them with > the restoredState. It's reasonable when a topic's partitions scaled. However, > if we add a new topic which has too much data and restore the Flink program, > the data of the new topic will be consumed from the start, which may not be > what we want. I think this should be an option for developers. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#issuecomment-513625214 ## CI report: * 37027083148923256920cbace9ef1f55b0bced5b : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119952271) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961 ## CI report: * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119766053) * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119948346) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner] [Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner] [Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190#issuecomment-513624385 Thanks a lot for your contribution to the Apache Flink project. I'm the @flinkbot. I help the community to review your pull request. We will use this comment to track the progress of the review. ## Review Progress * ❓ 1. The [description] looks good. * ❓ 2. There is [consensus] that the contribution should go into to Flink. * ❓ 3. Needs [attention] from. * ❓ 4. The change fits into the overall [architecture]. * ❓ 5. Overall code [quality] is good. Please see the [Pull Request Review Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full explanation of the review process. The Bot is tracking the review progress through labels. Labels are applied according to the order of the review items. For consensus, approval by a Flink committer of PMC member is required Bot commands The @flinkbot bot supports the following commands: - `@flinkbot approve description` to approve one or more aspects (aspects: `description`, `consensus`, `architecture` and `quality`) - `@flinkbot approve all` to approve all aspects - `@flinkbot approve-until architecture` to approve everything until `architecture` - `@flinkbot attention @username1 [@username2 ..]` to require somebody's attention - `@flinkbot disapprove architecture` to remove an approval you gave earlier This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
[ https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889884#comment-16889884 ] Liya Fan commented on FLINK-13221: -- [~ykt836]. Sure. Thanks a lot. > Blink planner should set ScheduleMode to > LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs > - > > Key: FLINK-13221 > URL: https://issues.apache.org/jira/browse/FLINK-13221 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Reporter: Kurt Young >Assignee: Liya Fan >Priority: Blocker > Labels: pull-request-available > Fix For: 1.9.0, 1.10.0 > > Time Spent: 20m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …
flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in … URL: https://github.com/apache/flink/pull/9165#issuecomment-512797712 ## CI report: * c51d878e9c6c9ccce6b1d67969aa029395acc4ea : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119628900) * a3ad9538892a23a689945f6cbf315625364dc756 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119951986) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] liyafan82 opened a new pull request #9190: [FLINK-13221][Table SQL / Planner] [Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batc
liyafan82 opened a new pull request #9190: [FLINK-13221][Table SQL / Planner] [Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9190 ## What is the purpose of the change Make batch jobs use LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST as the default scheduling mode. ## Brief change log - Replace LAZY_FROM_SOURCES with LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs. ## Verifying this change This change is already covered by existing tests, such as *(please describe tests)*. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Yarn/Mesos, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (not applicable) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] JingsongLi closed pull request #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink
JingsongLi closed pull request #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink URL: https://github.com/apache/flink/pull/9174 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink
JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink URL: https://github.com/apache/flink/pull/9174#issuecomment-513623907 I close this now, @wuchong will do some work to extract them to daily run. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink
JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink URL: https://github.com/apache/flink/pull/9174#issuecomment-513623792 Thanks @twalthr @wuchong , this tests should move to crontab and daily run. Every tests is important and precious. If we remove them, there maybe some bugs in future. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#issuecomment-510770597 ## CI report: * 6afb0645f386b618ddc1e9d86add2a22967676d6 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/118887349) * e5f48986b0ceb6bc774f33f3086898a5e6afb06c : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119123700) * a75172de1ff1d6c2a4b12dde013258fd7a36fc00 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119951727) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in … URL: https://github.com/apache/flink/pull/9165#discussion_r305660149 ## File path: flink-runtime/src/main/java/org/apache/flink/runtime/executiongraph/ExecutionGraphResultPartitionAvailabilityChecker.java ## @@ -0,0 +1,47 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.runtime.executiongraph; + +import org.apache.flink.runtime.executiongraph.failover.flip1.ResultPartitionAvailabilityChecker; +import org.apache.flink.runtime.io.network.partition.PartitionTracker; +import org.apache.flink.runtime.jobgraph.IntermediateResultPartitionID; + +import static org.apache.flink.util.Preconditions.checkNotNull; + +/** + * A {@link ResultPartitionAvailabilityChecker} which decides the intermediate result partition availability + * based on whether the corresponding result partition in the execution graph is tracked. + */ +public class ExecutionGraphResultPartitionAvailabilityChecker implements ResultPartitionAvailabilityChecker { + + /** The execution graph that hosts the result partitions. */ + private final ExecutionGraph executionGraph; + + /** The tracker that tracks all available result partitions. */ + private final PartitionTracker partitionTracker; + + ExecutionGraphResultPartitionAvailabilityChecker(final ExecutionGraph executionGraph) { Review comment: Good idea. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305659817 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/jobgraph/ScheduleModeTest.java ## @@ -31,6 +31,7 @@ @Test public void testAllowLazyDeployment() throws Exception { assertTrue(ScheduleMode.LAZY_FROM_SOURCES.allowLazyDeployment()); + assertTrue(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST.allowLazyDeployment()); Review comment: Reverted. Thanks. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305659829 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/minicluster/MiniClusterITCase.java ## @@ -117,7 +117,7 @@ public void testHandleStreamingJobsWhenNotEnoughSlot() throws Exception { @Test public void testHandleBatchJobsWhenNotEnoughSlot() throws Exception { try { - setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES); + setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST); Review comment: Reverted. Thanks. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305659842 ## File path: flink-streaming-java/src/test/java/org/apache/flink/streaming/api/graph/StreamingJobGraphGeneratorTest.java ## @@ -593,12 +593,19 @@ public void testDefaultScheduleMode() { public void testSetScheduleMode() { StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); - StreamGraph streamGraph = new StreamGraphGenerator(Collections.emptyList(), + StreamGraph streamGraph1 = new StreamGraphGenerator(Collections.emptyList(), env.getConfig(), env.getCheckpointConfig()) .setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES) .generate(); - JobGraph jobGraph = StreamingJobGraphGenerator.createJobGraph(streamGraph); - assertEquals(ScheduleMode.LAZY_FROM_SOURCES, jobGraph.getScheduleMode()); + JobGraph jobGraph1 = StreamingJobGraphGenerator.createJobGraph(streamGraph1); + assertEquals(ScheduleMode.LAZY_FROM_SOURCES, jobGraph1.getScheduleMode()); + + StreamGraph streamGraph2 = new StreamGraphGenerator(Collections.emptyList(), + env.getConfig(), env.getCheckpointConfig()) + .setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST) Review comment: Reverted. Thanks. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] liyafan82 commented on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
liyafan82 commented on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#issuecomment-513622448 > Thanks @liyafan82 for working on this. I think this change should stick to blink planner, we don't have to modify other runtime codes. @KurtYoung I see. Thanks a lot for your comments. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305659798 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java ## @@ -366,7 +366,7 @@ private JobGraph createBatchJobGraph() { v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, ResultPartitionType.BLOCKING); final JobGraph jobGraph = new JobGraph(v1, v2); - jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES); + jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST); Review comment: Reverted. Thanks. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in … URL: https://github.com/apache/flink/pull/9165#discussion_r305658754 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java ## @@ -379,10 +385,20 @@ private ExecutionGraph createExecutionGraph( final JobGraph jobGraph, final RestartStrategy restartStrategy) throws Exception { + final Configuration jmConfig = new Configuration(); + jmConfig.setBoolean(JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION, false); Review comment: Yes. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in … URL: https://github.com/apache/flink/pull/9165#discussion_r305658703 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGConcurrentFailoverTest.java ## @@ -250,37 +253,38 @@ public void testSkipFailoverIfExecutionStateIsNotRunning() throws Exception { */ private ExecutionGraph createExecutionGraph() throws Exception { - final JobInformation jobInformation = new DummyJobInformation(TEST_JOB_ID, "test job"); - final SimpleSlotProvider slotProvider = new SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM); - - final Time timeout = Time.seconds(10L); - final ExecutionGraph graph = new ExecutionGraph( - jobInformation, - TestingUtils.defaultExecutor(), - TestingUtils.defaultExecutor(), - timeout, - manuallyTriggeredRestartStrategy, - TestAdaptedRestartPipelinedRegionStrategyNG::new, - slotProvider, - getClass().getClassLoader(), - VoidBlobWriter.getInstance(), - timeout); - - JobVertex v1 = new JobVertex("vertex1"); + final JobVertex v1 = new JobVertex("vertex1"); v1.setInvokableClass(NoOpInvokable.class); v1.setParallelism(DEFAULT_PARALLELISM); - JobVertex v2 = new JobVertex("vertex2"); + final JobVertex v2 = new JobVertex("vertex2"); v2.setInvokableClass(NoOpInvokable.class); v2.setParallelism(DEFAULT_PARALLELISM); v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, ResultPartitionType.BLOCKING); - JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, v2); - graph.attachJobGraph(jg.getVerticesSortedTopologicallyFromSources()); + final JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, v2); + + final SimpleSlotProvider slotProvider = new SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM); + + final Configuration jmConfig = new Configuration(); + jmConfig.setBoolean(JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION, false); + + final PartitionTracker partitionTracker = new PartitionTrackerImpl( + jg.getJobID(), + NettyShuffleMaster.INSTANCE, + ignored -> Optional.empty()); + + final ExecutionGraph eg = new ExecutionGraphTestUtils.TestingExecutionGraphBuilder(jg) + .setRestartStrategy(manuallyTriggeredRestartStrategy) + .setFailoverStrategyFactory(TestAdaptedRestartPipelinedRegionStrategyNG::new) + .setSlotProvider(slotProvider) + .setJobMasterConfig(jmConfig) + .setPartitionTracker(partitionTracker) + .build(); - graph.start(componentMainThreadExecutor); + eg.start(componentMainThreadExecutor); - return graph; + return eg; Review comment: Ok. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in … URL: https://github.com/apache/flink/pull/9165#discussion_r305658771 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGConcurrentFailoverTest.java ## @@ -250,37 +253,38 @@ public void testSkipFailoverIfExecutionStateIsNotRunning() throws Exception { */ private ExecutionGraph createExecutionGraph() throws Exception { - final JobInformation jobInformation = new DummyJobInformation(TEST_JOB_ID, "test job"); - final SimpleSlotProvider slotProvider = new SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM); - - final Time timeout = Time.seconds(10L); - final ExecutionGraph graph = new ExecutionGraph( - jobInformation, - TestingUtils.defaultExecutor(), - TestingUtils.defaultExecutor(), - timeout, - manuallyTriggeredRestartStrategy, - TestAdaptedRestartPipelinedRegionStrategyNG::new, - slotProvider, - getClass().getClassLoader(), - VoidBlobWriter.getInstance(), - timeout); - - JobVertex v1 = new JobVertex("vertex1"); + final JobVertex v1 = new JobVertex("vertex1"); v1.setInvokableClass(NoOpInvokable.class); v1.setParallelism(DEFAULT_PARALLELISM); - JobVertex v2 = new JobVertex("vertex2"); + final JobVertex v2 = new JobVertex("vertex2"); v2.setInvokableClass(NoOpInvokable.class); v2.setParallelism(DEFAULT_PARALLELISM); v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, ResultPartitionType.BLOCKING); - JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, v2); - graph.attachJobGraph(jg.getVerticesSortedTopologicallyFromSources()); + final JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, v2); + + final SimpleSlotProvider slotProvider = new SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM); + + final Configuration jmConfig = new Configuration(); + jmConfig.setBoolean(JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION, false); Review comment: Yes. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner
lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner URL: https://github.com/apache/flink/pull/9181#issuecomment-513618686 @JingsongLi thanks for the review. One side note is blink planner requires a table source to implement `getReturnType()` method (hence the change to `HiveTableSource`). This method is marked as deprecated and seems the legacy planner doesn't require it. So I think it's better if we can avoid this inconsistency. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Closed] (FLINK-13329) Set env config for sql jobs
[ https://issues.apache.org/jira/browse/FLINK-13329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu closed FLINK-13329. --- Resolution: Won't Fix Fix Version/s: (was: 1.10.0) (was: 1.9.0) Closing this issue, because we don't need to solve this with discussed with [~julien1987]. > Set env config for sql jobs > --- > > Key: FLINK-13329 > URL: https://issues.apache.org/jira/browse/FLINK-13329 > Project: Flink > Issue Type: Task > Components: Table SQL / API >Affects Versions: 1.9.0, 1.10.0 >Reporter: XuPingyong >Priority: Major > > Now we execute streaming job through TableEnvironment, but > StreamExecutionEnvironment can not be touched by users, so we can not set > checkpoint and other env configs when we execute sql jobs. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-10806) Support multiple consuming offsets when discovering a new topic
[ https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889876#comment-16889876 ] Jiayi Liao commented on FLINK-10806: [~becket_qin] Firstly it's not only for a newly found partition but also newly topic. Assuming that I've already have a streaming pipeline for processing the data which are collected from user behaviours on web browsers. Now I want to add a new topic(user behaviours on mobile devices) into this pipeline, but this topic has existed for a while and already been used in other pipelines. So I stop the pipeline and add the topic into the KafkaConsumer topic lists and restart this pipeline, which causes that the data of this topic is consumed from earliest offset because it's a newly found TopicAndPartition. I think It's reasonable that we're able to consume from the latest offset of this newly topic. > Support multiple consuming offsets when discovering a new topic > --- > > Key: FLINK-10806 > URL: https://issues.apache.org/jira/browse/FLINK-10806 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka >Affects Versions: 1.6.2, 1.8.1 >Reporter: Jiayi Liao >Assignee: Jiayi Liao >Priority: Major > > In KafkaConsumerBase, we discover the TopicPartitions and compare them with > the restoredState. It's reasonable when a topic's partitions scaled. However, > if we add a new topic which has too much data and restore the Flink program, > the data of the new topic will be consumed from the start, which may not be > what we want. I think this should be an option for developers. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13341) Some streamTableSink#consumeDataStream(DataStream) don't be implemented as returning the sink transformation DataStreamSink when using blink planner.
[ https://issues.apache.org/jira/browse/FLINK-13341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13341: Priority: Critical (was: Major) > Some streamTableSink#consumeDataStream(DataStream) don't be implemented as > returning the sink transformation DataStreamSink when using blink planner. > - > > Key: FLINK-13341 > URL: https://issues.apache.org/jira/browse/FLINK-13341 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Reporter: chenqi >Priority: Critical > Labels: pull-request-available > Fix For: 1.9.0, 1.10.0 > > Time Spent: 10m > Remaining Estimate: 0h > > Some streamTableSink#consumeDataStream(DataStream) don't be implemented as > returning the sink transformation DataStreamSink when using blink planner. > which will throw the following errors: > {code:java} > Exception in thread "main" org.apache.flink.table.api.TableException: The > StreamTableSink#consumeDataStream(DataStream) must be implemented and return > the sink transformation DataStreamSink. However, > org.apache.flink.streaming.connectors.kafka.Kafka010TableSink doesn't > implement this method. > at > org.apache.flink.table.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:123) > at > org.apache.flink.table.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:50) > at > org.apache.flink.table.plan.nodes.exec.ExecNode$class.translateToPlan(ExecNode.scala:60) > at > org.apache.flink.table.plan.nodes.physical.stream.StreamExecSink.translateToPlan(StreamExecSink.scala:50) > at > org.apache.flink.table.planner.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:71) > at > org.apache.flink.table.planner.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:70) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) > at scala.collection.Iterator$class.foreach(Iterator.scala:891) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1334) > at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > at scala.collection.TraversableLike$class.map(TraversableLike.scala:234) > at scala.collection.AbstractTraversable.map(Traversable.scala:104) > at > org.apache.flink.table.planner.StreamPlanner.translateToPlan(StreamPlanner.scala:70) > at > org.apache.flink.table.planner.PlannerBase.translate(PlannerBase.scala:155) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:446) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:355) > at > org.apache.flink.table.examples.java.StreamSQLLookupJoinExample.main(StreamSQLLookupJoinExample.java:139){code} -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-12567) Rework DescriptorProperties to adapt unified DDL with clause and Descriptor key value pairs
[ https://issues.apache.org/jira/browse/FLINK-12567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-12567: Fix Version/s: (was: 1.9.0) 1.10.0 > Rework DescriptorProperties to adapt unified DDL with clause and Descriptor > key value pairs > --- > > Key: FLINK-12567 > URL: https://issues.apache.org/jira/browse/FLINK-12567 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Affects Versions: 1.8.0 >Reporter: Danny Chan >Assignee: Danny Chan >Priority: Major > Fix For: 1.10.0 > > > After introduce DDLs, we need to unify the k-v properties format for DDL with > clause and Descriptor API, this tool class also need to rework. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-12567) Rework DescriptorProperties to adapt unified DDL with clause and Descriptor key value pairs
[ https://issues.apache.org/jira/browse/FLINK-12567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-12567: Affects Version/s: (was: 1.8.0) > Rework DescriptorProperties to adapt unified DDL with clause and Descriptor > key value pairs > --- > > Key: FLINK-12567 > URL: https://issues.apache.org/jira/browse/FLINK-12567 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Reporter: Danny Chan >Assignee: Danny Chan >Priority: Major > Fix For: 1.10.0 > > > After introduce DDLs, we need to unify the k-v properties format for DDL with > clause and Descriptor API, this tool class also need to rework. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-12249) Type equivalence check fails for Window Aggregates
[ https://issues.apache.org/jira/browse/FLINK-12249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889874#comment-16889874 ] sunjincheng commented on FLINK-12249: - And I like the new update about workaround patch, in this way we can solve this issue for user in 1.9, and create a new JIRA for the follow-up discussion. just like [~dawidwys] and all of you mentioned this issue is not the blocker of 1.9 release. > Type equivalence check fails for Window Aggregates > -- > > Key: FLINK-12249 > URL: https://issues.apache.org/jira/browse/FLINK-12249 > Project: Flink > Issue Type: Bug > Components: Table SQL / Legacy Planner, Tests >Affects Versions: 1.9.0 >Reporter: Dawid Wysakowicz >Assignee: Hequn Cheng >Priority: Critical > Fix For: 1.9.0 > > > Creating Aggregate node fails in rules: {{LogicalWindowAggregateRule}} and > {{ExtendedAggregateExtractProjectRule}} if the only grouping expression is a > window and > we compute aggregation on NON NULLABLE field. > The root cause for that, is how return type inference strategies in calcite > work and how we handle window aggregates. Take > {{org.apache.calcite.sql.type.ReturnTypes#AGG_SUM}} as an example, based on > {{groupCount}} it adjusts type nullability based on groupCount. > Though we pass a false information as we strip down window aggregation from > groupSet (in {{LogicalWindowAggregateRule}}). > One can reproduce this problem also with a unit test like this: > {code} > @Test > def testTumbleFunction2() = { > > val innerQuery = > """ > |SELECT > | CASE a WHEN 1 THEN 1 ELSE 99 END AS correct, > | rowtime > |FROM MyTable > """.stripMargin > val sql = > "SELECT " + > " SUM(correct) as cnt, " + > " TUMBLE_START(rowtime, INTERVAL '15' MINUTE) as wStart " + > s"FROM ($innerQuery) " + > "GROUP BY TUMBLE(rowtime, INTERVAL '15' MINUTE)" > val expected = "" > streamUtil.verifySql(sql, expected) > } > {code} > This causes e2e tests to fail: > https://travis-ci.org/apache/flink/builds/521183361?utm_source=slack_medium=notificationhttps://travis-ci.org/apache/flink/builds/521183361?utm_source=slack_medium=notification -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-12557) Unify create table DDL with clause and connector descriptor keys
[ https://issues.apache.org/jira/browse/FLINK-12557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-12557: Fix Version/s: (was: 1.9.0) 1.10.0 > Unify create table DDL with clause and connector descriptor keys > > > Key: FLINK-12557 > URL: https://issues.apache.org/jira/browse/FLINK-12557 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Affects Versions: 1.8.0 >Reporter: Danny Chan >Assignee: Danny Chan >Priority: Major > Fix For: 1.10.0 > > > The *with* option in table DDL defines the properties needed for specific > connector to create TableSource/Sink. The properties structure for SqlClient > config YAML is defined in [Improvements to the Unified SQL Connector > API|https://docs.google.com/document/d/1Yaxp1UJUFW-peGLt8EIidwKIZEWrrA-pznWLuvaH39Y/edit#heading=h.41fd6rs7b3cf], > in this design, the properties can be categorized into 4 parts: > > # Top level properties: name, type(source, sink), update-mode ... > # Connector specific properties: connector.type, connector.path ... > # Format properties: format.type, format.fields.1.name ... > # Table schema properties: (can be omitted for DDL) > > This properties structure is reasonable for YAML, but they are not that > concise enough for developers. So there also defines a tool class named > [DescriptorProperties|https://github.com/apache/flink/blob/b3604f7bee7456b8533e9ea222a833a2624e36c2/flink-table/flink-table-common/src/main/java/org/apache/flink/table/descriptors/DescriptorProperties.java#L67] > to reconstruct the data structure(like TableSchema) from the flat k-v > strings. > > So in order to reduce complexity and keep the KV consistency for DDL with > properties and TableFactory properties, i proposed to simplify the DDL with > properties keys as following (corresponding to above 4 categories): > > # Top level properties: keep same as that in the YAML e.g. connector, > update-mode > # Connector specific properties: start with prefix named the connector type > e.g. for kafka connector, the properties are defined as kafka.k1 = v1, > kafka.k2 = v2 > # Format properties: format.type simplified to format and the others with > prefix of the format name e.g. format = 'json', json.line-delimiter = "\n" > # Table schema properties: omitted. > Here is a demo of creat table DDL: > {code:java} > CREATE TABLE Kafka10SourceTable ( > intField INTEGER, > stringField VARCHAR(128) COMMENT 'User IP address', > longField BIGINT, > rowTimeField TIMESTAMP, > WATERMARK wm01 FOR 'longField' AS BOUNDED WITH DELAY '60' SECOND > ) > COMMENT 'Kafka Source Table of topic user_ip_address' > WITH ( > connector='kafka', > kafka.property-version='1', > kafka.version='0.10', > kafka.topic='test-kafka-topic', > kafka.startup-mode = 'latest-offset' > kafka.specific-offset = 'offset' > format='json' > json.property-version = '1' > json.version='1' > json.derive-schema='true' > ) > {code} -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13105) Add documentation for blink planner's built-in functions
[ https://issues.apache.org/jira/browse/FLINK-13105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13105: Priority: Critical (was: Major) > Add documentation for blink planner's built-in functions > > > Key: FLINK-13105 > URL: https://issues.apache.org/jira/browse/FLINK-13105 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Planner >Reporter: Zhenghua Gao >Assignee: Zhenghua Gao >Priority: Critical > Labels: pull-request-available > Fix For: 1.9.0 > > Time Spent: 10m > Remaining Estimate: 0h > > Blink planner intros some built-in functions which need to be documented. -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-12557) Unify create table DDL with clause and connector descriptor keys
[ https://issues.apache.org/jira/browse/FLINK-12557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-12557: Affects Version/s: (was: 1.8.0) > Unify create table DDL with clause and connector descriptor keys > > > Key: FLINK-12557 > URL: https://issues.apache.org/jira/browse/FLINK-12557 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Reporter: Danny Chan >Assignee: Danny Chan >Priority: Major > Fix For: 1.10.0 > > > The *with* option in table DDL defines the properties needed for specific > connector to create TableSource/Sink. The properties structure for SqlClient > config YAML is defined in [Improvements to the Unified SQL Connector > API|https://docs.google.com/document/d/1Yaxp1UJUFW-peGLt8EIidwKIZEWrrA-pznWLuvaH39Y/edit#heading=h.41fd6rs7b3cf], > in this design, the properties can be categorized into 4 parts: > > # Top level properties: name, type(source, sink), update-mode ... > # Connector specific properties: connector.type, connector.path ... > # Format properties: format.type, format.fields.1.name ... > # Table schema properties: (can be omitted for DDL) > > This properties structure is reasonable for YAML, but they are not that > concise enough for developers. So there also defines a tool class named > [DescriptorProperties|https://github.com/apache/flink/blob/b3604f7bee7456b8533e9ea222a833a2624e36c2/flink-table/flink-table-common/src/main/java/org/apache/flink/table/descriptors/DescriptorProperties.java#L67] > to reconstruct the data structure(like TableSchema) from the flat k-v > strings. > > So in order to reduce complexity and keep the KV consistency for DDL with > properties and TableFactory properties, i proposed to simplify the DDL with > properties keys as following (corresponding to above 4 categories): > > # Top level properties: keep same as that in the YAML e.g. connector, > update-mode > # Connector specific properties: start with prefix named the connector type > e.g. for kafka connector, the properties are defined as kafka.k1 = v1, > kafka.k2 = v2 > # Format properties: format.type simplified to format and the others with > prefix of the format name e.g. format = 'json', json.line-delimiter = "\n" > # Table schema properties: omitted. > Here is a demo of creat table DDL: > {code:java} > CREATE TABLE Kafka10SourceTable ( > intField INTEGER, > stringField VARCHAR(128) COMMENT 'User IP address', > longField BIGINT, > rowTimeField TIMESTAMP, > WATERMARK wm01 FOR 'longField' AS BOUNDED WITH DELAY '60' SECOND > ) > COMMENT 'Kafka Source Table of topic user_ip_address' > WITH ( > connector='kafka', > kafka.property-version='1', > kafka.version='0.10', > kafka.topic='test-kafka-topic', > kafka.startup-mode = 'latest-offset' > kafka.specific-offset = 'offset' > format='json' > json.property-version = '1' > json.version='1' > json.derive-schema='true' > ) > {code} -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13300) Add OverwritableTableSink bridge logic to flink planner
[ https://issues.apache.org/jira/browse/FLINK-13300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13300: Fix Version/s: (was: 1.9.0) > Add OverwritableTableSink bridge logic to flink planner > - > > Key: FLINK-13300 > URL: https://issues.apache.org/jira/browse/FLINK-13300 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Affects Versions: 1.9.0 >Reporter: Danny Chan >Assignee: Danny Chan >Priority: Major > Fix For: 1.10.0 > > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13300) Add OverwritableTableSink bridge logic to flink planner
[ https://issues.apache.org/jira/browse/FLINK-13300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-13300: Affects Version/s: (was: 1.9.0) > Add OverwritableTableSink bridge logic to flink planner > - > > Key: FLINK-13300 > URL: https://issues.apache.org/jira/browse/FLINK-13300 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Reporter: Danny Chan >Assignee: Danny Chan >Priority: Major > Fix For: 1.10.0 > > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] sunjincheng121 commented on issue #9141: [FLINK-12249][table] Fix type equivalence check problems for Window Aggregates
sunjincheng121 commented on issue #9141: [FLINK-12249][table] Fix type equivalence check problems for Window Aggregates URL: https://github.com/apache/flink/pull/9141#issuecomment-513617339 Thanks for the update @hequn8128! I agree with @dawidwys @hequn8128 that this issue is not the blocker of 1.9 release. And I like this workaround patch, in this way we can solve this issue for user in 1.9, and create a new JIRA for the follow-up discussion. What do you think? Best, Jincheng This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (FLINK-13177) make FunctionCatalog able to reference functions across catalogs
[ https://issues.apache.org/jira/browse/FLINK-13177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889870#comment-16889870 ] Jark Wu commented on FLINK-13177: - Hi [~phoenixjiangnan], do we need this in 1.9.0? > make FunctionCatalog able to reference functions across catalogs > > > Key: FLINK-13177 > URL: https://issues.apache.org/jira/browse/FLINK-13177 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / API >Reporter: Bowen Li >Assignee: Bowen Li >Priority: Major > Fix For: 1.10.0 > > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-11433) JOIN on a table having a column of type MULTISET gives a NPE
[ https://issues.apache.org/jira/browse/FLINK-11433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-11433: Priority: Minor (was: Major) > JOIN on a table having a column of type MULTISET gives a NPE > - > > Key: FLINK-11433 > URL: https://issues.apache.org/jira/browse/FLINK-11433 > Project: Flink > Issue Type: Bug > Components: Table SQL / API >Affects Versions: 1.7.0, 1.7.1 >Reporter: Elias Saalmann >Assignee: TANG Wen-hui >Priority: Minor > > I get an error (Error while applying rule FlinkLogicalJoinConverter) when > performing a JOIN on a table having a column of type MULTISET (e.g. a COLLECT > as aggregation of a GROUP BY), for instance: > SELECT a, d > FROM TableA JOIN ( > SELECT b, COLLECT(c) AS d > FROM TableB > GROUP BY b > ) TableC ON a = b > Full stacktrace: > Exception in thread "main" java.lang.RuntimeException: Error while applying > rule FlinkLogicalJoinConverter, args > [rel#71:LogicalJoin.NONE(left=rel#69:Subset#3.NONE,right=rel#70:Subset#4.NONE,condition==($2, > $0),joinType=inner)] > at > org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:236) > at > org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:646) > at > org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:339) > at > org.apache.flink.table.api.TableEnvironment.runVolcanoPlanner(TableEnvironment.scala:373) > at > org.apache.flink.table.api.TableEnvironment.optimizeLogicalPlan(TableEnvironment.scala:292) > at > org.apache.flink.table.api.BatchTableEnvironment.optimize(BatchTableEnvironment.scala:455) > at > org.apache.flink.table.api.BatchTableEnvironment.translate(BatchTableEnvironment.scala:475) > at > org.apache.flink.table.api.java.BatchTableEnvironment.toDataSet(BatchTableEnvironment.scala:165) > at org.myorg.quickstart.TableJob2.main(TableJob2.java:40) > Caused by: java.lang.RuntimeException: Error occurred while applying rule > FlinkLogicalJoinConverter > at > org.apache.calcite.plan.volcano.VolcanoRuleCall.transformTo(VolcanoRuleCall.java:149) > at > org.apache.calcite.plan.RelOptRuleCall.transformTo(RelOptRuleCall.java:234) > at > org.apache.calcite.rel.convert.ConverterRule.onMatch(ConverterRule.java:141) > at > org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:212) > ... 8 more > Caused by: java.lang.NullPointerException > at > org.apache.flink.table.plan.nodes.FlinkRelNode$class.estimateDataTypeSize(FlinkRelNode.scala:84) > at > org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.estimateDataTypeSize(FlinkLogicalJoinBase.scala:29) > at > org.apache.flink.table.plan.nodes.FlinkRelNode$class.estimateDataTypeSize(FlinkRelNode.scala:104) > at > org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.estimateDataTypeSize(FlinkLogicalJoinBase.scala:29) > at > org.apache.flink.table.plan.nodes.FlinkRelNode$$anonfun$estimateRowSize$2.apply(FlinkRelNode.scala:80) > at > org.apache.flink.table.plan.nodes.FlinkRelNode$$anonfun$estimateRowSize$2.apply(FlinkRelNode.scala:79) > at > scala.collection.IndexedSeqOptimized$class.foldl(IndexedSeqOptimized.scala:57) > at > scala.collection.IndexedSeqOptimized$class.foldLeft(IndexedSeqOptimized.scala:66) > at scala.collection.mutable.ArrayBuffer.foldLeft(ArrayBuffer.scala:48) > at > org.apache.flink.table.plan.nodes.FlinkRelNode$class.estimateRowSize(FlinkRelNode.scala:79) > at > org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.estimateRowSize(FlinkLogicalJoinBase.scala:29) > at > org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.computeSelfCost(FlinkLogicalJoinBase.scala:48) > at > org.apache.calcite.rel.metadata.RelMdPercentageOriginalRows.getNonCumulativeCost(RelMdPercentageOriginalRows.java:162) > at > GeneratedMetadataHandler_NonCumulativeCost.getNonCumulativeCost_$(Unknown > Source) > at > GeneratedMetadataHandler_NonCumulativeCost.getNonCumulativeCost(Unknown > Source) > at > org.apache.calcite.rel.metadata.RelMetadataQuery.getNonCumulativeCost(RelMetadataQuery.java:301) > at > org.apache.calcite.plan.volcano.VolcanoPlanner.getCost(VolcanoPlanner.java:953) > at > org.apache.calcite.plan.volcano.RelSubset.propagateCostImprovements0(RelSubset.java:339) > at > org.apache.calcite.plan.volcano.RelSubset.propagateCostImprovements(RelSubset.java:322) > at > org.apache.calcite.plan.volcano.VolcanoPlanner.addRelToSet(VolcanoPlanner.java:1643) > at > org.apache.calcite.plan.volcano.VolcanoPlanner.registerImpl(VolcanoPlanner.java:1579) > at >
[GitHub] [flink] JingsongLi commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner
JingsongLi commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner URL: https://github.com/apache/flink/pull/9181#issuecomment-513613380 @lirui-apache Thanks for your contribution, LGTM This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961 ## CI report: * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE [Build](https://travis-ci.com/flink-ci/flink/builds/119766053) * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : PENDING [Build](https://travis-ci.com/flink-ci/flink/builds/119948346) This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Comment Edited] (FLINK-13231) Add a ratelimiter to pubsub source
[ https://issues.apache.org/jira/browse/FLINK-13231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889868#comment-16889868 ] Jiangjie Qin edited comment on FLINK-13231 at 7/22/19 1:54 AM: --- [~Xeli] Sorry for the late reply. The rate limiter is actually a per consumer / per sub-task rate limiter. I guess the Java doc is a little confusing, but that is actually what it is. was (Author: becket_qin): [~Xeli] The rate limiter is actually a per consumer / per sub-task rate limiter. I guess the Java doc is a little confusing, but that is actually what it is. > Add a ratelimiter to pubsub source > -- > > Key: FLINK-13231 > URL: https://issues.apache.org/jira/browse/FLINK-13231 > Project: Flink > Issue Type: Sub-task > Components: Connectors / Google Cloud PubSub >Reporter: Richard Deurwaarder >Priority: Minor > > Replace MaxMessagesToAcknowledge limit by introducing a rate limiter. See: > [https://github.com/apache/flink/pull/6594#discussion_r300215868] -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-13231) Add a ratelimiter to pubsub source
[ https://issues.apache.org/jira/browse/FLINK-13231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889868#comment-16889868 ] Jiangjie Qin commented on FLINK-13231: -- [~Xeli] The rate limiter is actually a per consumer / per sub-task rate limiter. I guess the Java doc is a little confusing, but that is actually what it is. > Add a ratelimiter to pubsub source > -- > > Key: FLINK-13231 > URL: https://issues.apache.org/jira/browse/FLINK-13231 > Project: Flink > Issue Type: Sub-task > Components: Connectors / Google Cloud PubSub >Reporter: Richard Deurwaarder >Priority: Minor > > Replace MaxMessagesToAcknowledge limit by introducing a rate limiter. See: > [https://github.com/apache/flink/pull/6594#discussion_r300215868] -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305650403 ## File path: flink-streaming-java/src/test/java/org/apache/flink/streaming/api/graph/StreamingJobGraphGeneratorTest.java ## @@ -593,12 +593,19 @@ public void testDefaultScheduleMode() { public void testSetScheduleMode() { StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); - StreamGraph streamGraph = new StreamGraphGenerator(Collections.emptyList(), + StreamGraph streamGraph1 = new StreamGraphGenerator(Collections.emptyList(), env.getConfig(), env.getCheckpointConfig()) .setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES) .generate(); - JobGraph jobGraph = StreamingJobGraphGenerator.createJobGraph(streamGraph); - assertEquals(ScheduleMode.LAZY_FROM_SOURCES, jobGraph.getScheduleMode()); + JobGraph jobGraph1 = StreamingJobGraphGenerator.createJobGraph(streamGraph1); + assertEquals(ScheduleMode.LAZY_FROM_SOURCES, jobGraph1.getScheduleMode()); + + StreamGraph streamGraph2 = new StreamGraphGenerator(Collections.emptyList(), + env.getConfig(), env.getCheckpointConfig()) + .setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST) Review comment: unnecessary changes? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305650383 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/minicluster/MiniClusterITCase.java ## @@ -117,7 +117,7 @@ public void testHandleStreamingJobsWhenNotEnoughSlot() throws Exception { @Test public void testHandleBatchJobsWhenNotEnoughSlot() throws Exception { try { - setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES); + setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST); Review comment: unnecessary changes? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305650367 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/jobgraph/ScheduleModeTest.java ## @@ -31,6 +31,7 @@ @Test public void testAllowLazyDeployment() throws Exception { assertTrue(ScheduleMode.LAZY_FROM_SOURCES.allowLazyDeployment()); + assertTrue(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST.allowLazyDeployment()); Review comment: unnecessary changes? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs URL: https://github.com/apache/flink/pull/9101#discussion_r305650341 ## File path: flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java ## @@ -366,7 +366,7 @@ private JobGraph createBatchJobGraph() { v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, ResultPartitionType.BLOCKING); final JobGraph jobGraph = new JobGraph(v1, v2); - jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES); + jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST); Review comment: unnecessary changes? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[jira] [Commented] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
[ https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889864#comment-16889864 ] Kurt Young commented on FLINK-13221: I changed this to 1.9.0's blocker, since this is a crucial one to fully turn on runtime's fine grained recovery for batch jobs. [~fan_li_ya] Could you also open a PR to 1.9 branch? > Blink planner should set ScheduleMode to > LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs > - > > Key: FLINK-13221 > URL: https://issues.apache.org/jira/browse/FLINK-13221 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Reporter: Kurt Young >Assignee: Liya Fan >Priority: Blocker > Labels: pull-request-available > Fix For: 1.9.0, 1.10.0 > > Time Spent: 10m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
[ https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kurt Young updated FLINK-13221: --- Fix Version/s: 1.10.0 1.9.0 > Blink planner should set ScheduleMode to > LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs > - > > Key: FLINK-13221 > URL: https://issues.apache.org/jira/browse/FLINK-13221 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Reporter: Kurt Young >Assignee: Liya Fan >Priority: Blocker > Labels: pull-request-available > Fix For: 1.9.0, 1.10.0 > > Time Spent: 10m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
[ https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kurt Young updated FLINK-13221: --- Priority: Blocker (was: Major) > Blink planner should set ScheduleMode to > LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs > - > > Key: FLINK-13221 > URL: https://issues.apache.org/jira/browse/FLINK-13221 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Reporter: Kurt Young >Assignee: Liya Fan >Priority: Blocker > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Updated] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
[ https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kurt Young updated FLINK-13221: --- Issue Type: Bug (was: Improvement) > Blink planner should set ScheduleMode to > LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs > - > > Key: FLINK-13221 > URL: https://issues.apache.org/jira/browse/FLINK-13221 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner >Reporter: Kurt Young >Assignee: Liya Fan >Priority: Major > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[jira] [Commented] (FLINK-13199) ARM support for Flink
[ https://issues.apache.org/jira/browse/FLINK-13199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889861#comment-16889861 ] wangxiyuan commented on FLINK-13199: Hi Chesnay: It's similar. The difference is that which CI system will be used. OpenLab has two ways to support ARM build for Apache community. One is that Openlab provides CI system directly, like travis-ci, Flink can add `openlab-ci`[1] as the ARM build CI the same as travis-ci for X86. The other ways is what is said on Infra ML, Openlab can donate ARM resources to Infra team, and then Flink can use the apache offical CI system[2] as the ARM build CI. For the first way, OpenLab team can help to write CI jobs scripts and keep maintaining the related CI jobs. For the second way, I guess Infra team will handle it. [1]: [https://status.openlabtesting.org/status] [2]: [https://builds.apache.org/] > ARM support for Flink > - > > Key: FLINK-13199 > URL: https://issues.apache.org/jira/browse/FLINK-13199 > Project: Flink > Issue Type: New Feature > Components: Build System >Reporter: wangxiyuan >Priority: Critical > > There is not official ARM release for Flink. But basing on my local test, > Flink which is made by Java and Scala is built and tested well. So is it > possible to support ARM release officially? And I think it's may not be a > huge work. > > AFAIK, Flink now uses travis-ci which supports only x86 for CI gate. Is it > possible to add an ARM one? I'm from openlab community[1]. Similar with > travis-ci, it's is an opensource and free community which provide CI > resources and system for opensource projects, contains both ARM and X86 > machines. And now it helps some community building there CI already. Such as > OpenStack and CNCF. > > If Flink community agree to support ARM. I can spend my full time to help. > Such as job define, CI maintaining, test fix and so on. If Flink don't want > to rely on OpenLab, we can donate ARM resources directly as well. > > I have sent out a discuess mail-list already[2]. Feel free to reply there or > here. > > Thanks. > > [1]:[https://openlabtesting.org/] > [2]:[http://mail-archives.apache.org/mod_mbox/flink-dev/201907.mbox/browser] -- This message was sent by Atlassian JIRA (v7.6.14#76016)
[GitHub] [flink] TsReaper commented on a change in pull request #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, f
TsReaper commented on a change in pull request #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serializer restore in BaseArray/Map serializer and add tests for complex data formats URL: https://github.com/apache/flink/pull/9139#discussion_r305649009 ## File path: flink-core/src/test/java/org/apache/flink/api/common/typeutils/SerializerTestBase.java ## @@ -124,30 +124,8 @@ public void testSnapshotConfigurationAndReconfigure() throws Exception { final TypeSerializer serializer = getSerializer(); final TypeSerializerSnapshot configSnapshot = serializer.snapshotConfiguration(); - byte[] serializedConfig; - try (ByteArrayOutputStream out = new ByteArrayOutputStream()) { - TypeSerializerSnapshotSerializationUtil.writeSerializerSnapshot( - new DataOutputViewStreamWrapper(out), configSnapshot, serializer); - serializedConfig = out.toByteArray(); - } - - TypeSerializerSnapshot restoredConfig; - try (ByteArrayInputStream in = new ByteArrayInputStream(serializedConfig)) { - restoredConfig = TypeSerializerSnapshotSerializationUtil.readSerializerSnapshot( - new DataInputViewStreamWrapper(in), Thread.currentThread().getContextClassLoader(), getSerializer()); - } - - TypeSerializerSchemaCompatibility strategy = restoredConfig.resolveSchemaCompatibility(getSerializer()); - final TypeSerializer restoreSerializer; - if (strategy.isCompatibleAsIs()) { - restoreSerializer = restoredConfig.restoreSerializer(); - } - else if (strategy.isCompatibleWithReconfiguredSerializer()) { - restoreSerializer = strategy.getReconfiguredSerializer(); - } - else { - throw new AssertionError("Unable to restore serializer with " + strategy); - } + final TypeSerializer restoreSerializer = snapshotAndReconfigure( Review comment: No. I just move these into a separate method so that it can be reused for my newly added tests. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] sjwiesman edited a comment on issue #9161: [FLINK-13262][docs] Add documentation for the new Table & SQL API type system
sjwiesman edited a comment on issue #9161: [FLINK-13262][docs] Add documentation for the new Table & SQL API type system URL: https://github.com/apache/flink/pull/9161#issuecomment-513228516 @twalthr Looks good to me +1 This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services
[GitHub] [flink] tweise commented on issue #9187: [FLINK-12595][kinesis] Interrupt thread at right time to avoid deadlock
tweise commented on issue #9187: [FLINK-12595][kinesis] Interrupt thread at right time to avoid deadlock URL: https://github.com/apache/flink/pull/9187#issuecomment-513580704 @flinkbot approve all This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services