[GitHub] [flink] flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default 
partition name of Hive table
URL: https://github.com/apache/flink/pull/9088#issuecomment-510484364
 
 
   ## CI report:
   
   * eac5f74690ddb0b08cb41b029f5b8ac675e63565 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/118891245)
   * b22f836f5e8f95a9f376f54c68798eeb14cb1644 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119054589)
   * 64007bc344772aa3496f5d8b0a456f73466bbd17 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119417375)
   * efd6e3189c7ee347335c0ec028603caea2001740 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441225)
   * d8b308866574c18cdcd6c89cda7f5837ef50352d : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119955505)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9191: [FLINK-11767] Introduce new 
TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
URL: https://github.com/apache/flink/pull/9191#issuecomment-513649015
 
 
   ## CI report:
   
   * ddd758fdc5714f75114c1cd35cd61e45091cf736 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119960105)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] 
Add CachedMiniCluster to share cluster between ITCases
URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961
 
 
   ## CI report:
   
   * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766053)
   * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119948346)
   * 7536e2fb5a510a0ca4dfb46476a79307011e9286 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119954049)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase

2019-07-21 Thread GitBox
flinkbot commented on issue #9191: [FLINK-11767] Introduce new 
TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
URL: https://github.com/apache/flink/pull/9191#issuecomment-513649015
 
 
   ## CI report:
   
   * ddd758fdc5714f75114c1cd35cd61e45091cf736 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119960105)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase

2019-07-21 Thread GitBox
flinkbot commented on issue #9191: [FLINK-11767] Introduce new 
TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
URL: https://github.com/apache/flink/pull/9191#issuecomment-513647351
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-11767) Update TypeSerializerSnapshotMigrationTestBase and subclasses for 1.8

2019-07-21 Thread ASF GitHub Bot (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-11767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-11767:
---
Labels: pull-request-available  (was: )

> Update TypeSerializerSnapshotMigrationTestBase and subclasses for 1.8
> -
>
> Key: FLINK-11767
> URL: https://issues.apache.org/jira/browse/FLINK-11767
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Type Serialization System, Tests
>Affects Versions: 1.8.0
>Reporter: vinoyang
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>
> Update {{TypeSerializerSnapshotMigrationTestBase}} and subclasses to cover 
> restoring from Flink 1.8.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-10806) Support multiple consuming offsets when discovering a new topic

2019-07-21 Thread Jiayi Liao (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889932#comment-16889932
 ] 

Jiayi Liao commented on FLINK-10806:


[~becket_qin] Thanks for pointing this out. What you described is more like a 
migration of state(migrate T1 state into Pipeline B), which reminds me that the 
savepoint connector that community is building now, can't support the state 
migration of kafka(maybe you can help build it ^_^). 

I've already support the simple strategy(earliest/latest) in our own flink 
branch. But your concerns make sense so I can't give a conclusion now. Maybe we 
can wait until we get more user cases and scenarios?

> Support multiple consuming offsets when discovering a new topic
> ---
>
> Key: FLINK-10806
> URL: https://issues.apache.org/jira/browse/FLINK-10806
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.6.2, 1.8.1
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Major
>
> In KafkaConsumerBase, we discover the TopicPartitions and compare them with 
> the restoredState. It's reasonable when a topic's partitions scaled. However, 
> if we add a new topic which has too much data and restore the Flink program, 
> the data of the new topic will be consumed from the start, which may not be 
> what we want.  I think this should be an option for developers.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] tzulitai opened a new pull request #9191: [FLINK-11767] Introduce new TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase

2019-07-21 Thread GitBox
tzulitai opened a new pull request #9191: [FLINK-11767] Introduce new 
TypeSerializerUpgradeTestBase to replace TypeSerializerSnapshotMigrationTestBase
URL: https://github.com/apache/flink/pull/9191
 
 
   #What is the purpose of the change
   
   This PR introduces a new `TypeSerializerUprgadeTestBase` to replace the 
older `TypeSerializerSnapshotMigrationTestBase`. The new test base allows 
subclasses to define different versions of the same runtime class, by 
relocating different classes to be under the same name using the 
`ClassRelocator`.
   
   This PR already tries the new test base out with the `PojoSerializer`.
   
   ## Verifying this change
   
   This is a test related change.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? n/a


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9190: [FLINK-13221][Table SQL / 
Planner][Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#issuecomment-513625214
 
 
   ## CI report:
   
   * 37027083148923256920cbace9ef1f55b0bced5b : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119952271)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-13094) Provide an easy way to read timers using the State Processor API

2019-07-21 Thread Tzu-Li (Gordon) Tai (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13094?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai closed FLINK-13094.
---
   Resolution: Fixed
Fix Version/s: 1.9.0

Merged.

master (1.10.0): 752e7ae6a43455234606a9a7da50f40182b2db51
1.9.0: b8eec2b822af8648c73422bcb86ae6b3822420a9

> Provide an easy way to read timers using the State Processor API
> 
>
> Key: FLINK-13094
> URL: https://issues.apache.org/jira/browse/FLINK-13094
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream, Runtime / State Backends
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Seth Wiesman
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13293) Add state processor api to opt/ directory in flink-dist

2019-07-21 Thread Tzu-Li (Gordon) Tai (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai updated FLINK-13293:

Affects Version/s: (was: 1.10.0)
   (was: 1.9.0)

> Add state processor api to opt/ directory in flink-dist
> ---
>
> Key: FLINK-13293
> URL: https://issues.apache.org/jira/browse/FLINK-13293
> Project: Flink
>  Issue Type: Bug
>  Components: Build System
>Reporter: Seth Wiesman
>Assignee: Seth Wiesman
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Closed] (FLINK-13293) Add state processor api to opt/ directory in flink-dist

2019-07-21 Thread Tzu-Li (Gordon) Tai (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai closed FLINK-13293.
---
   Resolution: Fixed
Fix Version/s: (was: 1.10.0)

Merged.

master (1.10.0): c8af3c95dbac76458c01bd50e4d6f85aeca3e89a
1.9.0: 752e7ae6a43455234606a9a7da50f40182b2db51

> Add state processor api to opt/ directory in flink-dist
> ---
>
> Key: FLINK-13293
> URL: https://issues.apache.org/jira/browse/FLINK-13293
> Project: Flink
>  Issue Type: Bug
>  Components: Build System
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Seth Wiesman
>Assignee: Seth Wiesman
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM 
tracked partitions to find out unavailable partitions in …
URL: https://github.com/apache/flink/pull/9165#issuecomment-512797712
 
 
   ## CI report:
   
   * c51d878e9c6c9ccce6b1d67969aa029395acc4ea : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119628900)
   * a3ad9538892a23a689945f6cbf315625364dc756 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119951986)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9088: [FLINK-13012][hive] Handle default 
partition name of Hive table
URL: https://github.com/apache/flink/pull/9088#issuecomment-510484364
 
 
   ## CI report:
   
   * eac5f74690ddb0b08cb41b029f5b8ac675e63565 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/118891245)
   * b22f836f5e8f95a9f376f54c68798eeb14cb1644 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119054589)
   * 64007bc344772aa3496f5d8b0a456f73466bbd17 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119417375)
   * efd6e3189c7ee347335c0ec028603caea2001740 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119441225)
   * d8b308866574c18cdcd6c89cda7f5837ef50352d : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119955505)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] 
Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#issuecomment-510770597
 
 
   ## CI report:
   
   * 6afb0645f386b618ddc1e9d86add2a22967676d6 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/118887349)
   * e5f48986b0ceb6bc774f33f3086898a5e6afb06c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119123700)
   * a75172de1ff1d6c2a4b12dde013258fd7a36fc00 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119951727)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serialize

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9139: 
[FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation 
of getString and getBinary method in NestedRow, fix serializer restore in 
BaseArray/Map serializer and add tests for complex data formats
URL: https://github.com/apache/flink/pull/9139#issuecomment-512166499
 
 
   ## CI report:
   
   * 22609f1b6271176affec70926b89b4730451568a : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119436577)
   * 7dbec40e5321c5ffcaae89c8716851a8a1248519 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119572205)
   * 85740dd1f663191b172b2601c2e4a0a4c5167ad1 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119625173)
   * eeb0b5cea4c0a633e76ef61667bdb52d6bad520f : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119636253)
   * 3519a3632830d4edc2b90a1cc2b0e467ebcb823c : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766271)
   * 8e1e152e94f0114237baf148e8e9052afb9c2756 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119953465)
   * a2a42395a583f1268de18486ee0beef6eeb491cc : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119956594)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] asfgit closed pull request #9094: [FLINK-13094][state-processor-api] Provide an easy way to read timers using the State Processor API

2019-07-21 Thread GitBox
asfgit closed pull request #9094: [FLINK-13094][state-processor-api] Provide an 
easy way to read timers using the State Processor API
URL: https://github.com/apache/flink/pull/9094
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections

2019-07-21 Thread TisonKun (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889905#comment-16889905
 ] 

TisonKun commented on FLINK-10052:
--

Yes. LeaderContender(jm, rm, dispatcher, etc.) in FLINK is all subclass of 
{{FencedRpcEndpoint}}, which means that rpc messages of different leadership 
epoch are fenced.

> Tolerate temporarily suspended ZooKeeper connections
> 
>
> Key: FLINK-10052
> URL: https://issues.apache.org/jira/browse/FLINK-10052
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1
>Reporter: Till Rohrmann
>Assignee: Dominik Wosiński
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This issue results from FLINK-10011 which uncovered a problem with Flink's HA 
> recovery and proposed the following solution to harden Flink:
> The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator 
> recipe for leader election. The leader latch revokes leadership in case of a 
> suspended ZooKeeper connection. This can be premature in case that the system 
> can reconnect to ZooKeeper before its session expires. The effect of the lost 
> leadership is that all jobs will be canceled and directly restarted after 
> regaining the leadership.
> Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper 
> connection, it would be better to wait until the ZooKeeper connection is 
> LOST. That way we would allow the system to reconnect and not lose the 
> leadership. This could be achievable by using Curator's {{LeaderSelector}} 
> instead of the {{LeaderLatch}}.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] asfgit closed pull request #9133: [FLINK-13293][state-processor-api][build] Add state processor api to opt/ directory in flink-dist

2019-07-21 Thread GitBox
asfgit closed pull request #9133: [FLINK-13293][state-processor-api][build] Add 
state processor api to opt/ directory in flink-dist
URL: https://github.com/apache/flink/pull/9133
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] tzulitai commented on issue #9094: [FLINK-13094][state-processor-api] Provide an easy way to read timers using the State Processor API

2019-07-21 Thread GitBox
tzulitai commented on issue #9094: [FLINK-13094][state-processor-api] Provide 
an easy way to read timers using the State Processor API
URL: https://github.com/apache/flink/pull/9094#issuecomment-513636515
 
 
   Merging ...


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] tzulitai commented on issue #9133: [FLINK-13293][state-processor-api][build] Add state processor api to opt/ directory in flink-dist

2019-07-21 Thread GitBox
tzulitai commented on issue #9133: [FLINK-13293][state-processor-api][build] 
Add state processor api to opt/ directory in flink-dist
URL: https://github.com/apache/flink/pull/9133#issuecomment-513636024
 
 
   +1, merging ...


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-13346) indexdb reporter parse data type error

2019-07-21 Thread Yun Tang (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13346?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yun Tang closed FLINK-13346.

Resolution: Duplicate

> indexdb reporter parse data type error
> --
>
> Key: FLINK-13346
> URL: https://issues.apache.org/jira/browse/FLINK-13346
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.7.2, 1.8.0, 1.8.1
> Environment: Yarn cluster
>Reporter: Jake.zhang
>Priority: Minor
>
> unable to parse 
> 'taskmanager_job_task_operator_KafkaConsumer_sync-time-max,host=dn45,job_id=f9bc38f9d31ac689715c055a58905288,job_name=network_monitor_save_phoenix,operator_id=cbc357ccb763df2852fee8c4fc7d55f2,operator_name=Source:\
>  Custom\ 
> Source,subtask_index=0,task_attempt_id=3462928b2995a24303eefad4ebfb9f4f,task_attempt_num=0,task_id=cbc357ccb763df2852fee8c4fc7d55f2,task_name=Source:\
>  Custom\ Source\ ->\ Map\ ->\ Filter\ ->\ Sink:\ 
> Unnamed,tm_id=container_e65_1560932557384_276950_01_000647 value=-∞ 
> 156376155872700': invalid number dropped=0
>  at 
> org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionFromErrorMessage(InfluxDBException.java:147)
>  at 
> org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionForErrorState(InfluxDBException.java:173)
>  at 
> org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.execute(InfluxDBImpl.java:796)
>  at 
> org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.write(InfluxDBImpl.java:455)
>  at 
> org.apache.flink.metrics.influxdb.InfluxdbReporter.report(InfluxdbReporter.java:97)
>  at 
> org.apache.flink.runtime.metrics.MetricRegistryImpl$ReporterTask.run(MetricRegistryImpl.java:427)
>  at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>  at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections

2019-07-21 Thread Quan Shi (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889901#comment-16889901
 ] 

Quan Shi commented on FLINK-10052:
--

hi [~Tison]

Sounds reasonable. Elect a new leader and recovery from checkpoints is a 
concise way for HA.  What happens to FLINK if zookeeper split-brain or 
zookeeper cluster shutdown? Does FLINK has fencing mechanism like HDFS? 

> Tolerate temporarily suspended ZooKeeper connections
> 
>
> Key: FLINK-10052
> URL: https://issues.apache.org/jira/browse/FLINK-10052
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1
>Reporter: Till Rohrmann
>Assignee: Dominik Wosiński
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This issue results from FLINK-10011 which uncovered a problem with Flink's HA 
> recovery and proposed the following solution to harden Flink:
> The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator 
> recipe for leader election. The leader latch revokes leadership in case of a 
> suspended ZooKeeper connection. This can be premature in case that the system 
> can reconnect to ZooKeeper before its session expires. The effect of the lost 
> leadership is that all jobs will be canceled and directly restarted after 
> regaining the leadership.
> Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper 
> connection, it would be better to wait until the ZooKeeper connection is 
> LOST. That way we would allow the system to reconnect and not lose the 
> leadership. This could be achievable by using Curator's {{LeaderSelector}} 
> instead of the {{LeaderLatch}}.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] lirui-apache commented on issue #9088: [FLINK-13012][hive] Handle default partition name of Hive table

2019-07-21 Thread GitBox
lirui-apache commented on issue #9088: [FLINK-13012][hive] Handle default 
partition name of Hive table
URL: https://github.com/apache/flink/pull/9088#issuecomment-513633987
 
 
   Changed to test with table env. This PR needs to wait until we have #9181 
merged.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serialize

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9139: 
[FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation 
of getString and getBinary method in NestedRow, fix serializer restore in 
BaseArray/Map serializer and add tests for complex data formats
URL: https://github.com/apache/flink/pull/9139#issuecomment-512166499
 
 
   ## CI report:
   
   * 22609f1b6271176affec70926b89b4730451568a : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119436577)
   * 7dbec40e5321c5ffcaae89c8716851a8a1248519 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119572205)
   * 85740dd1f663191b172b2601c2e4a0a4c5167ad1 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119625173)
   * eeb0b5cea4c0a633e76ef61667bdb52d6bad520f : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119636253)
   * 3519a3632830d4edc2b90a1cc2b0e467ebcb823c : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766271)
   * 8e1e152e94f0114237baf148e8e9052afb9c2756 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119953465)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13344) Translate "How to Contribute" page into Chinese.

2019-07-21 Thread Jeff Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889895#comment-16889895
 ] 

Jeff Yang commented on FLINK-13344:
---

I got it . Thanks [~jark]

> Translate "How to Contribute" page into Chinese. 
> -
>
> Key: FLINK-13344
> URL: https://issues.apache.org/jira/browse/FLINK-13344
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: WangHengWei
>Priority: Major
>
> The page is https://flink.apache.org/zh/contributing/how-to-contribute.html
> The markdown file is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md
> Before start working on this, please read translation guideline: 
> https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13344) Translate "How to Contribute" page into Chinese.

2019-07-21 Thread WangHengWei (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889894#comment-16889894
 ] 

WangHengWei commented on FLINK-13344:
-

[~jark], thanks very much. I'm gonna do it.:D

> Translate "How to Contribute" page into Chinese. 
> -
>
> Key: FLINK-13344
> URL: https://issues.apache.org/jira/browse/FLINK-13344
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: WangHengWei
>Priority: Major
>
> The page is https://flink.apache.org/zh/contributing/how-to-contribute.html
> The markdown file is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md
> Before start working on this, please read translation guideline: 
> https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections

2019-07-21 Thread TisonKun (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889893#comment-16889893
 ] 

TisonKun commented on FLINK-10052:
--

[~quan]

Supposed you have jm-1 and jm-2 execute the same job but run on different 
cluster. jm-1 is the previous leader and thus jm-2 is inactive. Now jm-1 lost 
connection with zookeeper and jm-2 became the leader, recovering from job graph 
store and checkpoint store, started executing the job. If we didn't shutdown 
jm-1 and supposed it would be shutdown due to zk communication failure, not 
only there were two jms running for the same job thus the same job was executed 
twice, but also they might conflict writing states.

To sum up, if a jm lost connection with zk, there is possibly another jm 
running for the same job. We don't want to see 2 jms running for the same job 
at the same time, right?

If you're worry about the performance, when you executing a streaming job, 
flink can recover it from checkpoints and catch up quickly.

> Tolerate temporarily suspended ZooKeeper connections
> 
>
> Key: FLINK-10052
> URL: https://issues.apache.org/jira/browse/FLINK-10052
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1
>Reporter: Till Rohrmann
>Assignee: Dominik Wosiński
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This issue results from FLINK-10011 which uncovered a problem with Flink's HA 
> recovery and proposed the following solution to harden Flink:
> The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator 
> recipe for leader election. The leader latch revokes leadership in case of a 
> suspended ZooKeeper connection. This can be premature in case that the system 
> can reconnect to ZooKeeper before its session expires. The effect of the lost 
> leadership is that all jobs will be canceled and directly restarted after 
> regaining the leadership.
> Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper 
> connection, it would be better to wait until the ZooKeeper connection is 
> LOST. That way we would allow the system to reconnect and not lose the 
> leadership. This could be achievable by using Curator's {{LeaderSelector}} 
> instead of the {{LeaderLatch}}.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] 
Add CachedMiniCluster to share cluster between ITCases
URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961
 
 
   ## CI report:
   
   * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766053)
   * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119948346)
   * 7536e2fb5a510a0ca4dfb46476a79307011e9286 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119954049)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread GitBox
KurtYoung commented on issue #9190: [FLINK-13221][Table SQL / 
Planner][Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#issuecomment-51363
 
 
   LGTM, waiting travis pass


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUE

2019-07-21 Thread GitBox
KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL 
/ Planner][Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#discussion_r305665394
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java
 ##
 @@ -455,6 +455,15 @@ protected void restartTasks(final Set 
verticesToRestart) {
this.delegate = checkNotNull(delegate);
}
 
+   @Override
+   public CompletableFuture allocateBatchSlot(
 
 Review comment:
   Thanks for the clarification!


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "

2019-07-21 Thread Jingsong Lee (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-13289:
-
Priority: Blocker  (was: Major)

> Blink Planner JDBCUpsertTableSink : UnsupportedOperationException 
> "JDBCUpsertTableSink can not support "
> 
>
> Key: FLINK-13289
> URL: https://issues.apache.org/jira/browse/FLINK-13289
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC, Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: LakeShen
>Priority: Blocker
> Fix For: 1.9.0, 1.10.0
>
>
> Hi , in flink-jdbc connector module, I change the Flink planner to Blink 
> planner to test all test case,because we want to use Blank planner in our 
> program. When I test the JDBCUpsertTableSinkITCase class , the method 
> testUpsert throw the exception:
> {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can 
> not support {color}
> I saw the src code,in Flink planner , the StreamPlanner set the 
> JDBCUpsertTableSink' keyFields,
> but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' 
> keyFields,so JDBCUpsertTableSink keyFields is null, when execute 
> JDBCUpsertTableSink newFormat(),
> it thrown the exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13342) Flink1.9 document number error

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13342?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13342:

Priority: Critical  (was: Major)

> Flink1.9 document number error
> --
>
> Key: FLINK-13342
> URL: https://issues.apache.org/jira/browse/FLINK-13342
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.9.0
>Reporter: hanfei
>Priority: Critical
> Fix For: 1.9.0
>
>
> link:https://flink.apache.org/
> description:Document menu flink1.9(Snapshot) jump to flink1.10(Snapshot)
> bug:Flink1.9 document number error



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Comment Edited] (FLINK-13242) StandaloneResourceManagerTest fails on travis

2019-07-21 Thread Xintong Song (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13242?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889630#comment-16889630
 ] 

Xintong Song edited comment on FLINK-13242 at 7/22/19 3:45 AM:
---

Hi [~azagrebin], I think I found the problem.

In _StandaloneResourceManager#initialize()_ it uses _getMainThreadExecutor()_ 
to execute _setFailUnfulfillableRequest()_. However, before 
_setFailUnfulfillableRequest()_ is executed, the main thread executor of the 
resource manager might be replaced by a new one when it accepts the granted 
leader ship, leading to _setFailUnfulfillableRequest()_ never being executed. 
This only happens when the _StandaloneResourceManager#initialize()_ is invoked 
before _TestingLeaderElectionService#isLeader()_.

The problem can be re-produced and verified as follows:
 * Add logs in _StandaloneResourceManager#initialize()_, 
_TestingLeaderElectionService#isLeader()_ and 
_ResourceManager#setFailUnfulfillableRequest()_, and run the test. In most 
cases, you should see _TestingLeaderElectionService#isLeader()_ invoked before 
_StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is 
invoked, and the test should pass.
 * Add a short sleep time (in my case 100ms) in 
_MockResourceManagerRuntimeServices#grantLeadership()_ before 
_rmLeaderElectionService.isLeader()_, and run the test again. Now you should 
see _TestingLeaderElectionService#isLeader()_ invoked after 
_StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is 
never invoked, and the test should fail.
 * Add another short sleep time (also 100ms in my case) in 
_StandaloneResourceManager#initialize()_, inside the 
_getRpcService().getScheduledExecutor().schedule()_ block, right before 
_getMainThreadExecutor()_. This should change the order of invoking back and 
fix the failure.
 * If you invoke _getMainThreadExecutor()_ twice in 
_StandaloneResourceManager#initialize()_, once before the sleep and the other 
after it, and print out the fetched main thread executors, you should find that 
they are two different objects.
 * Now if you remove the sleep in _StandaloneResourceManager#initialize()_, you 
should see the printed two main thread executors are the same object, and the 
test is broken again.

I'm thinking that maybe _setFailUnfulfillableRequest(true)_ does not need to be 
invoked on the PRC main thread. Instead of calling on the main thread executor, 
I tried call _setFailUnfulfillableRequest(true)_ directly in the 
_getRpcService().getScheduledExecutor().schedule()_ block in 
_StandaloneResourceManager#initialize()_ and it fixes the problem.

I think we do not care whether the _setFailUnfulfillableRequest(true)_ happens 
on main thread or not in production, as long as it eventually get invoked. And 
for this test case, we may have a bit inconsistency that after 
_setFailUnfulfillableRequest(true)_ the _isFailingUnfulfillableRequest()_ may 
not get the correct result immediately, which I think is acceptable and the 10s 
timeout for _assertHappensUntil()_ should be long enough to catch the invoking 
of _setFailUnfulfillableRequest(true)_ eventually. What do you think?


was (Author: xintongsong):
Hi [~azagrebin], I think I found the problem.

In _StandaloneResourceManager#initialize()_ it uses _getMainThreadExecutor()_ 
to execute _setFailUnfulfillableRequest()_. However, before 
_setFailUnfulfillableRequest()_ is executed, the main thread executor of the 
resource manager might be replaced by a new one when it accepts the granted 
leader ship, leading to _setFailUnfulfillableRequest()_ never being executed. 
This only happens when the _StandaloneResourceManager#initialize()_ is invoked 
before _TestingLeaderElectionService#isLeader()_.

The problem can be re-produced and verified as follows:
 * Add logs in _StandaloneResourceManager#initialize()_, __ 
_TestingLeaderElectionService#isLeader()_ and 
_ResourceManager#setFailUnfulfillableRequest()_, and run the test. In most 
cases, you should see _TestingLeaderElectionService#isLeader()_ invoked before 
__ _StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is 
invoked, and the test should pass.
 * Add a short sleep time (in my case 100ms) in 
_MockResourceManagerRuntimeServices#grantLeadership()_ before 
_rmLeaderElectionService.isLeader()_, and run the test again. Now you should 
see _TestingLeaderElectionService#isLeader()_ invoked after __ 
_StandaloneResourceManager#initialize()_, _setFailUnfulfillableRequest()_ is 
never invoked, and the test should fail.
 * Add another short sleep time (also 100ms in my case) in 
_StandaloneResourceManager#initialize()_, inside __ the 
_getRpcService().getScheduledExecutor().schedule()_ block, right before 
_getMainThreadExecutor()_. This should change the order of invoking back and 
fix the failure.
 * If you invoke _getMainThreadExecutor()_ twice in 

[jira] [Commented] (FLINK-13318) Blink planner tests failing on Scala 2.12

2019-07-21 Thread Jark Wu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889891#comment-16889891
 ] 

Jark Wu commented on FLINK-13318:
-

+1 for this approach!

> Blink planner tests failing on Scala 2.12
> -
>
> Key: FLINK-13318
> URL: https://issues.apache.org/jira/browse/FLINK-13318
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner, Tests
>Affects Versions: 1.9.0
>Reporter: Chesnay Schepler
>Priority: Blocker
> Fix For: 1.9.0
>
>
> [https://travis-ci.org/apache/flink/builds/559909681]
> {code:java}
> 13:30:03.531 [INFO] Results:
> 13:30:03.531 [INFO] 
> 13:30:03.533 [ERROR] Failures: 
> 13:30:03.534 [ERROR]   CalcTest.testScalarFunctionAccess:64 planBefore 
> expected:<...t$giveMeCaseClass$$f[e1bff2b06d8e0e495536102224cfe83().my], 
> _c1=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$fe1bff2b06d8e0e495536102224cfe83().clazz],
>  
> _c2=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$fe1bff2b06d8e0e495536102224cfe83().my],
>  
> _c3=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$fe1bff2b06d8e0e495536102224cfe83]().clazz])
> +- Logica...> but 
> was:<...t$giveMeCaseClass$$f[4a420732fc04b1351889eb0e88eb891().my], 
> _c1=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$f4a420732fc04b1351889eb0e88eb891().clazz],
>  
> _c2=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$f4a420732fc04b1351889eb0e88eb891().my],
>  
> _c3=[org$apache$flink$table$plan$batch$table$CalcTest$giveMeCaseClass$$f4a420732fc04b1351889eb0e88eb891]().clazz])
> +- Logica...>
> 13:30:03.534 [ERROR]   CalcTest.testSelectFromGroupedTableWithFunctionKey:154 
> planBefore 
> expected:<...alcTest$MyHashCode$$[d14b486109d9dd062ae7c60e0497797]5($2)])
>   +- Log...> but 
> was:<...alcTest$MyHashCode$$[3cd929923219fc59162b13a4941ead4]5($2)])
>   +- Log...>
> 13:30:03.534 [ERROR]   CalcTest.testSelectFunction:109 planBefore 
> expected:<...alcTest$MyHashCode$$[d14b486109d9dd062ae7c60e0497797]5($2)], 
> b=[$1])
> +- L...> but 
> was:<...alcTest$MyHashCode$$[3cd929923219fc59162b13a4941ead4]5($2)], b=[$1])
> +- L...>
> 13:30:03.534 [ERROR]   CorrelateTest.testCrossJoin:41 planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], 
> rowType=[Rec...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], 
> rowType=[Rec...>
> 13:30:03.534 [ERROR]   CorrelateTest.testCrossJoin2:52 planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2, 
> _UTF-16LE'$')]...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2, 
> _UTF-16LE'$')]...>
> 13:30:03.534 [ERROR]   CorrelateTest.testLeftOuterJoinWithLiteralTrue:74 
> planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], 
> rowType=[Rec...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], 
> rowType=[Rec...>
> 13:30:03.534 [ERROR]   
> CorrelateTest.testLeftOuterJoinWithoutJoinPredicates:63 planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], 
> rowType=[Rec...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], 
> rowType=[Rec...>
> 13:30:03.535 [ERROR]   JoinTest.testFilterJoinRule:143 planBefore 
> expected:<...le$JoinTest$Merger$$[223b7380fec29c4077a893c60165d845($2, 
> org$apache$flink$table$plan$batch$table$JoinTest$Merger$$223b7380fec29c4077a893c60165d845]($2,
>  $5))])
>+- Lo...> but 
> was:<...le$JoinTest$Merger$$[d18a3011491fab359eccb50f2d0d9a18($2, 
> org$apache$flink$table$plan$batch$table$JoinTest$Merger$$d18a3011491fab359eccb50f2d0d9a18]($2,
>  $5))])
>+- Lo...>
> 13:30:03.535 [ERROR]   CorrelateStringExpressionTest.testCorrelateJoins1:39 
> planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], 
> rowType=[Rec...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], 
> rowType=[Rec...>
> 13:30:03.535 [ERROR]   CorrelateStringExpressionTest.testCorrelateJoins2:45 
> planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2)], 
> rowType=[Rec...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2)], 
> rowType=[Rec...>
> 13:30:03.535 [ERROR]   CorrelateStringExpressionTest.testCorrelateJoins3:51 
> planBefore 
> expected:<...ble$util$TableFunc1$[ad38060966060e704b09fa4c9428769]6($2, 
> _UTF-16LE'$')]...> but 
> was:<...ble$util$TableFunc1$[e1a0c63ecf595c7329d87aae4f6f425]6($2, 
> _UTF-16LE'$')]...>
> 13:30:03.535 [ERROR]   CorrelateStringExpressionTest.testCorrelateJoins4:57 
> planBefore 
> expected:<...ble$util$TableFunc2$[b3b1f988779be024ed9386bce5019112]($2)], 
> rowType=[Reco...> but 
> 

[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13289:

Fix Version/s: 1.10.0

> Blink Planner JDBCUpsertTableSink : UnsupportedOperationException 
> "JDBCUpsertTableSink can not support "
> 
>
> Key: FLINK-13289
> URL: https://issues.apache.org/jira/browse/FLINK-13289
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC, Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: LakeShen
>Priority: Major
> Fix For: 1.9.0, 1.10.0
>
>
> Hi , in flink-jdbc connector module, I change the Flink planner to Blink 
> planner to test all test case,because we want to use Blank planner in our 
> program. When I test the JDBCUpsertTableSinkITCase class , the method 
> testUpsert throw the exception:
> {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can 
> not support {color}
> I saw the src code,in Flink planner , the StreamPlanner set the 
> JDBCUpsertTableSink' keyFields,
> but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' 
> keyFields,so JDBCUpsertTableSink keyFields is null, when execute 
> JDBCUpsertTableSink newFormat(),
> it thrown the exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13289:

Fix Version/s: 1.9.0

> Blink Planner JDBCUpsertTableSink : UnsupportedOperationException 
> "JDBCUpsertTableSink can not support "
> 
>
> Key: FLINK-13289
> URL: https://issues.apache.org/jira/browse/FLINK-13289
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC, Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: LakeShen
>Priority: Major
> Fix For: 1.9.0
>
>
> Hi , in flink-jdbc connector module, I change the Flink planner to Blink 
> planner to test all test case,because we want to use Blank planner in our 
> program. When I test the JDBCUpsertTableSinkITCase class , the method 
> testUpsert throw the exception:
> {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can 
> not support {color}
> I saw the src code,in Flink planner , the StreamPlanner set the 
> JDBCUpsertTableSink' keyFields,
> but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' 
> keyFields,so JDBCUpsertTableSink keyFields is null, when execute 
> JDBCUpsertTableSink newFormat(),
> it thrown the exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13289:

Component/s: Connectors / JDBC

> Blink Planner JDBCUpsertTableSink : UnsupportedOperationException 
> "JDBCUpsertTableSink can not support "
> 
>
> Key: FLINK-13289
> URL: https://issues.apache.org/jira/browse/FLINK-13289
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC, Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: LakeShen
>Priority: Major
>
> Hi , in flink-jdbc connector module, I change the Flink planner to Blink 
> planner to test all test case,because we want to use Blank planner in our 
> program. When I test the JDBCUpsertTableSinkITCase class , the method 
> testUpsert throw the exception:
> {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can 
> not support {color}
> I saw the src code,in Flink planner , the StreamPlanner set the 
> JDBCUpsertTableSink' keyFields,
> but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' 
> keyFields,so JDBCUpsertTableSink keyFields is null, when execute 
> JDBCUpsertTableSink newFormat(),
> it thrown the exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13289) Blink Planner JDBCUpsertTableSink : UnsupportedOperationException "JDBCUpsertTableSink can not support "

2019-07-21 Thread Jark Wu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889890#comment-16889890
 ] 

Jark Wu commented on FLINK-13289:
-

Thanks [~kakachen] for digging into this. This is definitely a bug in blink 
planner. We will fix this in planner.

> Blink Planner JDBCUpsertTableSink : UnsupportedOperationException 
> "JDBCUpsertTableSink can not support "
> 
>
> Key: FLINK-13289
> URL: https://issues.apache.org/jira/browse/FLINK-13289
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: LakeShen
>Priority: Major
>
> Hi , in flink-jdbc connector module, I change the Flink planner to Blink 
> planner to test all test case,because we want to use Blank planner in our 
> program. When I test the JDBCUpsertTableSinkITCase class , the method 
> testUpsert throw the exception:
> {color:red}java.lang.UnsupportedOperationException: JDBCUpsertTableSink can 
> not support {color}
> I saw the src code,in Flink planner , the StreamPlanner set the 
> JDBCUpsertTableSink' keyFields,
> but in Blink planner , I didn't find anywhere to set JDBCUpsertTableSink' 
> keyFields,so JDBCUpsertTableSink keyFields is null, when execute 
> JDBCUpsertTableSink newFormat(),
> it thrown the exception.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, fix serialize

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9139: 
[FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation 
of getString and getBinary method in NestedRow, fix serializer restore in 
BaseArray/Map serializer and add tests for complex data formats
URL: https://github.com/apache/flink/pull/9139#issuecomment-512166499
 
 
   ## CI report:
   
   * 22609f1b6271176affec70926b89b4730451568a : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119436577)
   * 7dbec40e5321c5ffcaae89c8716851a8a1248519 : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119572205)
   * 85740dd1f663191b172b2601c2e4a0a4c5167ad1 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119625173)
   * eeb0b5cea4c0a633e76ef61667bdb52d6bad520f : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119636253)
   * 3519a3632830d4edc2b90a1cc2b0e467ebcb823c : SUCCESS 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766271)
   * 8e1e152e94f0114237baf148e8e9052afb9c2756 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119953465)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] godfreyhe commented on a change in pull request #9056: [FLINK-13185] [sql-parser][table-planner] Bump Calcite dependency to 1.20.0 in sql parser & flink planner

2019-07-21 Thread GitBox
godfreyhe commented on a change in pull request #9056: [FLINK-13185] 
[sql-parser][table-planner] Bump Calcite dependency to 1.20.0 in sql parser & 
flink planner
URL: https://github.com/apache/flink/pull/9056#discussion_r305664376
 
 

 ##
 File path: 
flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/plan/nodes/CommonCorrelate.scala
 ##
 @@ -46,7 +47,7 @@ trait CommonCorrelate {
 inputSchema: RowSchema,
 udtfTypeInfo: TypeInformation[Any],
 returnSchema: RowSchema,
-joinType: SemiJoinType,
+joinType: JoinRelType,
 
 Review comment:
   sorry for late reply, I create a JIRA to fix this: 
https://issues.apache.org/jira/browse/FLINK-13347


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-13347) should handle new JoinRelType(SEMI/ANTI) in switch case

2019-07-21 Thread godfrey he (JIRA)
godfrey he created FLINK-13347:
--

 Summary: should handle new JoinRelType(SEMI/ANTI) in switch case
 Key: FLINK-13347
 URL: https://issues.apache.org/jira/browse/FLINK-13347
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Legacy Planner, Table SQL / Planner
Reporter: godfrey he
 Fix For: 1.9.0, 1.10.0


Calcite 1.20 introduces {{SEMI}} & {{ANTI}} to {{JoinRelType}}, blink planner & 
flink planner should handle them in each switch case



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13344) Translate "How to Contribute" page into Chinese.

2019-07-21 Thread Jark Wu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889888#comment-16889888
 ] 

Jark Wu commented on FLINK-13344:
-

Hi [~highfei2...@126.com], this is not a duplicate issue with FLINK-13343, this 
is another page :)
Hi [~WangHW], I assigned this issue to you. 

> Translate "How to Contribute" page into Chinese. 
> -
>
> Key: FLINK-13344
> URL: https://issues.apache.org/jira/browse/FLINK-13344
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: WangHengWei
>Priority: Major
>
> The page is https://flink.apache.org/zh/contributing/how-to-contribute.html
> The markdown file is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md
> Before start working on this, please read translation guideline: 
> https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Assigned] (FLINK-13344) Translate "How to Contribute" page into Chinese.

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-13344:
---

Assignee: WangHengWei

> Translate "How to Contribute" page into Chinese. 
> -
>
> Key: FLINK-13344
> URL: https://issues.apache.org/jira/browse/FLINK-13344
> Project: Flink
>  Issue Type: Sub-task
>  Components: chinese-translation, Project Website
>Reporter: Jark Wu
>Assignee: WangHengWei
>Priority: Major
>
> The page is https://flink.apache.org/zh/contributing/how-to-contribute.html
> The markdown file is located in 
> https://github.com/apache/flink-web/blob/asf-site/contributing/how-to-contribute.zh.md
> Before start working on this, please read translation guideline: 
> https://cwiki.apache.org/confluence/display/FLINK/Flink+Translation+Specifications



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-10052) Tolerate temporarily suspended ZooKeeper connections

2019-07-21 Thread Quan Shi (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889887#comment-16889887
 ] 

Quan Shi commented on FLINK-10052:
--

Hi, all. 
What if the Job Manager works fine except losing connection with zookeeper? The 
stream job can continue running. Do we really need to select another active Job 
Manager and restart the whole job?

> Tolerate temporarily suspended ZooKeeper connections
> 
>
> Key: FLINK-10052
> URL: https://issues.apache.org/jira/browse/FLINK-10052
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.4.2, 1.5.2, 1.6.0, 1.8.1
>Reporter: Till Rohrmann
>Assignee: Dominik Wosiński
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This issue results from FLINK-10011 which uncovered a problem with Flink's HA 
> recovery and proposed the following solution to harden Flink:
> The {{ZooKeeperLeaderElectionService}} uses the {{LeaderLatch}} Curator 
> recipe for leader election. The leader latch revokes leadership in case of a 
> suspended ZooKeeper connection. This can be premature in case that the system 
> can reconnect to ZooKeeper before its session expires. The effect of the lost 
> leadership is that all jobs will be canceled and directly restarted after 
> regaining the leadership.
> Instead of directly revoking the leadership upon a SUSPENDED ZooKeeper 
> connection, it would be better to wait until the ZooKeeper connection is 
> LOST. That way we would allow the system to reconnect and not lose the 
> leadership. This could be achievable by using Curator's {{LeaderSelector}} 
> instead of the {{LeaderLatch}}.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] liyafan82 commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUE

2019-07-21 Thread GitBox
liyafan82 commented on a change in pull request #9190: [FLINK-13221][Table SQL 
/ Planner][Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#discussion_r305663679
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java
 ##
 @@ -455,6 +455,15 @@ protected void restartTasks(final Set 
verticesToRestart) {
this.delegate = checkNotNull(delegate);
}
 
+   @Override
+   public CompletableFuture allocateBatchSlot(
 
 Review comment:
   Yes. Without this method override, the default implementation of the super 
class will be called, which is based on SlotProvider#allocateBatchSlot. The 
default implementation throws an UnsupportedOperationException, which will 
cause some UT to fail.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUE

2019-07-21 Thread GitBox
KurtYoung commented on a change in pull request #9190: [FLINK-13221][Table SQL 
/ Planner][Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#discussion_r305663260
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java
 ##
 @@ -455,6 +455,15 @@ protected void restartTasks(final Set 
verticesToRestart) {
this.delegate = checkNotNull(delegate);
}
 
+   @Override
+   public CompletableFuture allocateBatchSlot(
 
 Review comment:
   Is this necessary?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-13346) indexdb reporter parse data type error

2019-07-21 Thread Jake.zhang (JIRA)
Jake.zhang created FLINK-13346:
--

 Summary: indexdb reporter parse data type error
 Key: FLINK-13346
 URL: https://issues.apache.org/jira/browse/FLINK-13346
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.8.1, 1.8.0, 1.7.2
 Environment: Yarn cluster
Reporter: Jake.zhang


unable to parse 
'taskmanager_job_task_operator_KafkaConsumer_sync-time-max,host=dn45,job_id=f9bc38f9d31ac689715c055a58905288,job_name=network_monitor_save_phoenix,operator_id=cbc357ccb763df2852fee8c4fc7d55f2,operator_name=Source:\
 Custom\ 
Source,subtask_index=0,task_attempt_id=3462928b2995a24303eefad4ebfb9f4f,task_attempt_num=0,task_id=cbc357ccb763df2852fee8c4fc7d55f2,task_name=Source:\
 Custom\ Source\ ->\ Map\ ->\ Filter\ ->\ Sink:\ 
Unnamed,tm_id=container_e65_1560932557384_276950_01_000647 value=-∞ 
156376155872700': invalid number dropped=0
 at 
org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionFromErrorMessage(InfluxDBException.java:147)
 at 
org.apache.flink.metrics.influxdb.shaded.org.influxdb.InfluxDBException.buildExceptionForErrorState(InfluxDBException.java:173)
 at 
org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.execute(InfluxDBImpl.java:796)
 at 
org.apache.flink.metrics.influxdb.shaded.org.influxdb.impl.InfluxDBImpl.write(InfluxDBImpl.java:455)
 at 
org.apache.flink.metrics.influxdb.InfluxdbReporter.report(InfluxdbReporter.java:97)
 at 
org.apache.flink.runtime.metrics.MetricRegistryImpl$ReporterTask.run(MetricRegistryImpl.java:427)
 at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
 at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
 at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
 at java.lang.Thread.run(Thread.java:745)



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] JingsongLi commented on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases

2019-07-21 Thread GitBox
JingsongLi commented on issue #9180: [FLINK-13331][table-planner-blink] Add 
CachedMiniCluster to share cluster between ITCases
URL: https://github.com/apache/flink/pull/9180#issuecomment-513625785
 
 
   @wuchong @zentol Thanks for your attention. 
   
   About leak resources, there is a clean thread to invoke 
`MiniClusterWithClientResource.after` to clean resources. Compared with the 
past, it only delayed cleaning up resources by 15 seconds.
   
   >I have doubts whether this will improve the test times in a significant 
way, since we're only skipping the creation of ~60 mini clusters.
   
   Yeah, we should gauge how valuable this change is.
   1.I test on my macbook pro, blink-planner will use 17min before, and use 
11min now for `mvn clean verify`. If we just run `mvn verify`, the gap will be 
more pronounced.
   2.But the more accurate test time should be travis triggering, I will update 
later.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-10806) Support multiple consuming offsets when discovering a new topic

2019-07-21 Thread Jiangjie Qin (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889886#comment-16889886
 ] 

Jiangjie Qin commented on FLINK-10806:
--

[~wind_ljy] Thanks for the context. Yes, I agree that the use case you 
mentioned is a pretty reasonable use case. And the fact currently the newly 
discovered partitions (they may or may not belong to a new topic) are set to 
{{EARLIEST_OFFSET}} is a bad behavior.

What I am wondering is whether you only need to set the consumer offset to the 
latest for newly discovered partitions, or you may need to set the offset to 
some specific offset. This may impact the delivery guarantee.

For example, consider the following event sequence:
 # Initial state: Pipeline A is processing Topic T1. Pipeline B is processing 
Topic T2. For simplicity, let's assume topic T1 and T2 both only have 1 
partition (P0).
 # Stop pipeline A and pipeline B. Reassign both topic 1 and topic 2 to 
Pipeline B.
 # The upstream producers produced a few more messages into T1.
 # Restart pipeline B to consume from both T1 and T2.

In the above case, after restart pipeline B, for T2 the consumption will start 
from the last committed offset, which is probably what we expect. However, for 
T1, if the consumption starts from the latest, the messages produced in step 3 
will be lost. Is that behavior expected? Do you want to set the starting offset 
to where pipeline A left in this case?

It is quite easy to just always allow users to configure the consumer to either 
start from earliest or latest for the new partitions if the group offset is 
missing. But allowing a specific starting offset for a newly discovered topic 
needs a few more thoughts.

> Support multiple consuming offsets when discovering a new topic
> ---
>
> Key: FLINK-10806
> URL: https://issues.apache.org/jira/browse/FLINK-10806
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.6.2, 1.8.1
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Major
>
> In KafkaConsumerBase, we discover the TopicPartitions and compare them with 
> the restoredState. It's reasonable when a topic's partitions scaled. However, 
> if we add a new topic which has too much data and restore the Flink program, 
> the data of the new topic will be consumed from the start, which may not be 
> what we want.  I think this should be an option for developers.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner][Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread GitBox
flinkbot commented on issue #9190: [FLINK-13221][Table SQL / 
Planner][Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#issuecomment-513625214
 
 
   ## CI report:
   
   * 37027083148923256920cbace9ef1f55b0bced5b : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119952271)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] 
Add CachedMiniCluster to share cluster between ITCases
URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961
 
 
   ## CI report:
   
   * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766053)
   * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119948346)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner] [Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread GitBox
flinkbot commented on issue #9190: [FLINK-13221][Table SQL / Planner] 
[Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190#issuecomment-513624385
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread Liya Fan (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889884#comment-16889884
 ] 

Liya Fan commented on FLINK-13221:
--

[~ykt836]. Sure. Thanks a lot.

> Blink planner should set ScheduleMode to 
> LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
> -
>
> Key: FLINK-13221
> URL: https://issues.apache.org/jira/browse/FLINK-13221
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Kurt Young
>Assignee: Liya Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0, 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9165: [FLINK-13055] [runtime] Leverage JM 
tracked partitions to find out unavailable partitions in …
URL: https://github.com/apache/flink/pull/9165#issuecomment-512797712
 
 
   ## CI report:
   
   * c51d878e9c6c9ccce6b1d67969aa029395acc4ea : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119628900)
   * a3ad9538892a23a689945f6cbf315625364dc756 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119951986)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liyafan82 opened a new pull request #9190: [FLINK-13221][Table SQL / Planner] [Release-1.9] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batc

2019-07-21 Thread GitBox
liyafan82 opened a new pull request #9190: [FLINK-13221][Table SQL / Planner] 
[Release-1.9] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9190
 
 
   
   
   ## What is the purpose of the change
   
   Make batch jobs use LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST as the default 
scheduling mode.
   
   
   ## Brief change log
   
 - Replace LAZY_FROM_SOURCES with LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST 
for batch jobs.
 
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] JingsongLi closed pull request #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink

2019-07-21 Thread GitBox
JingsongLi closed pull request #9174: [FLINK-13330][table-planner-blink] Remove 
unnecessary to reduce testing time in blink
URL: https://github.com/apache/flink/pull/9174
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink

2019-07-21 Thread GitBox
JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove 
unnecessary to reduce testing time in blink
URL: https://github.com/apache/flink/pull/9174#issuecomment-513623907
 
 
   I close this now, @wuchong will do some work to extract them to daily run.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove unnecessary to reduce testing time in blink

2019-07-21 Thread GitBox
JingsongLi commented on issue #9174: [FLINK-13330][table-planner-blink] Remove 
unnecessary to reduce testing time in blink
URL: https://github.com/apache/flink/pull/9174#issuecomment-513623792
 
 
   Thanks @twalthr @wuchong , this tests should move to crontab and daily run. 
Every tests is important and precious. If we remove them, there maybe some bugs 
in future.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9101: [FLINK-13221][Table SQL / Planner] 
Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#issuecomment-510770597
 
 
   ## CI report:
   
   * 6afb0645f386b618ddc1e9d86add2a22967676d6 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/118887349)
   * e5f48986b0ceb6bc774f33f3086898a5e6afb06c : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119123700)
   * a75172de1ff1d6c2a4b12dde013258fd7a36fc00 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119951727)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …

2019-07-21 Thread GitBox
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] 
Leverage JM tracked partitions to find out unavailable partitions in …
URL: https://github.com/apache/flink/pull/9165#discussion_r305660149
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/executiongraph/ExecutionGraphResultPartitionAvailabilityChecker.java
 ##
 @@ -0,0 +1,47 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.executiongraph;
+
+import 
org.apache.flink.runtime.executiongraph.failover.flip1.ResultPartitionAvailabilityChecker;
+import org.apache.flink.runtime.io.network.partition.PartitionTracker;
+import org.apache.flink.runtime.jobgraph.IntermediateResultPartitionID;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * A {@link ResultPartitionAvailabilityChecker} which decides the intermediate 
result partition availability
+ * based on whether the corresponding result partition in the execution graph 
is tracked.
+ */
+public class ExecutionGraphResultPartitionAvailabilityChecker implements 
ResultPartitionAvailabilityChecker {
+
+   /** The execution graph that hosts the result partitions. */
+   private final ExecutionGraph executionGraph;
+
+   /** The tracker that tracks all available result partitions. */
+   private final PartitionTracker partitionTracker;
+
+   ExecutionGraphResultPartitionAvailabilityChecker(final ExecutionGraph 
executionGraph) {
 
 Review comment:
   Good idea.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305659817
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/jobgraph/ScheduleModeTest.java
 ##
 @@ -31,6 +31,7 @@
@Test
public void testAllowLazyDeployment() throws Exception {

assertTrue(ScheduleMode.LAZY_FROM_SOURCES.allowLazyDeployment());
+   
assertTrue(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST.allowLazyDeployment());
 
 Review comment:
   Reverted. Thanks.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305659829
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/minicluster/MiniClusterITCase.java
 ##
 @@ -117,7 +117,7 @@ public void testHandleStreamingJobsWhenNotEnoughSlot() 
throws Exception {
@Test
public void testHandleBatchJobsWhenNotEnoughSlot() throws Exception {
try {
-   
setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES);
+   
setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST);
 
 Review comment:
   Reverted. Thanks.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305659842
 
 

 ##
 File path: 
flink-streaming-java/src/test/java/org/apache/flink/streaming/api/graph/StreamingJobGraphGeneratorTest.java
 ##
 @@ -593,12 +593,19 @@ public void testDefaultScheduleMode() {
public void testSetScheduleMode() {
StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 
-   StreamGraph streamGraph = new 
StreamGraphGenerator(Collections.emptyList(),
+   StreamGraph streamGraph1 = new 
StreamGraphGenerator(Collections.emptyList(),
env.getConfig(), env.getCheckpointConfig())
.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES)
.generate();
-   JobGraph jobGraph = 
StreamingJobGraphGenerator.createJobGraph(streamGraph);
-   assertEquals(ScheduleMode.LAZY_FROM_SOURCES, 
jobGraph.getScheduleMode());
+   JobGraph jobGraph1 = 
StreamingJobGraphGenerator.createJobGraph(streamGraph1);
+   assertEquals(ScheduleMode.LAZY_FROM_SOURCES, 
jobGraph1.getScheduleMode());
+
+   StreamGraph streamGraph2 = new 
StreamGraphGenerator(Collections.emptyList(),
+   env.getConfig(), env.getCheckpointConfig())
+   
.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST)
 
 Review comment:
   Reverted. Thanks.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liyafan82 commented on issue #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread GitBox
liyafan82 commented on issue #9101: [FLINK-13221][Table SQL / Planner] Blink 
planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST 
for batch jobs
URL: https://github.com/apache/flink/pull/9101#issuecomment-513622448
 
 
   > Thanks @liyafan82 for working on this. I think this change should stick to 
blink planner, we don't have to modify other runtime codes.
   
   @KurtYoung I see. Thanks a lot for your comments.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
liyafan82 commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305659798
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java
 ##
 @@ -366,7 +366,7 @@ private JobGraph createBatchJobGraph() {
v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, 
ResultPartitionType.BLOCKING);
 
final JobGraph jobGraph = new JobGraph(v1, v2);
-   jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES);
+   
jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST);
 
 Review comment:
   Reverted. Thanks.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …

2019-07-21 Thread GitBox
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] 
Leverage JM tracked partitions to find out unavailable partitions in …
URL: https://github.com/apache/flink/pull/9165#discussion_r305658754
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java
 ##
 @@ -379,10 +385,20 @@ private ExecutionGraph createExecutionGraph(
final JobGraph jobGraph,
final RestartStrategy restartStrategy) throws Exception 
{
 
+   final Configuration jmConfig = new Configuration();
+   
jmConfig.setBoolean(JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION, 
false);
 
 Review comment:
   Yes.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …

2019-07-21 Thread GitBox
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] 
Leverage JM tracked partitions to find out unavailable partitions in …
URL: https://github.com/apache/flink/pull/9165#discussion_r305658703
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGConcurrentFailoverTest.java
 ##
 @@ -250,37 +253,38 @@ public void 
testSkipFailoverIfExecutionStateIsNotRunning() throws Exception {
 */
private ExecutionGraph createExecutionGraph() throws Exception {
 
-   final JobInformation jobInformation = new 
DummyJobInformation(TEST_JOB_ID, "test job");
-   final SimpleSlotProvider slotProvider = new 
SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM);
-
-   final Time timeout = Time.seconds(10L);
-   final ExecutionGraph graph = new ExecutionGraph(
-   jobInformation,
-   TestingUtils.defaultExecutor(),
-   TestingUtils.defaultExecutor(),
-   timeout,
-   manuallyTriggeredRestartStrategy,
-   TestAdaptedRestartPipelinedRegionStrategyNG::new,
-   slotProvider,
-   getClass().getClassLoader(),
-   VoidBlobWriter.getInstance(),
-   timeout);
-
-   JobVertex v1 = new JobVertex("vertex1");
+   final JobVertex v1 = new JobVertex("vertex1");
v1.setInvokableClass(NoOpInvokable.class);
v1.setParallelism(DEFAULT_PARALLELISM);
 
-   JobVertex v2 = new JobVertex("vertex2");
+   final JobVertex v2 = new JobVertex("vertex2");
v2.setInvokableClass(NoOpInvokable.class);
v2.setParallelism(DEFAULT_PARALLELISM);
 
v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, 
ResultPartitionType.BLOCKING);
 
-   JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, v2);
-   
graph.attachJobGraph(jg.getVerticesSortedTopologicallyFromSources());
+   final JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, 
v2);
+
+   final SimpleSlotProvider slotProvider = new 
SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM);
+
+   final Configuration jmConfig = new Configuration();
+   
jmConfig.setBoolean(JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION, 
false);
+
+   final PartitionTracker partitionTracker = new 
PartitionTrackerImpl(
+   jg.getJobID(),
+   NettyShuffleMaster.INSTANCE,
+   ignored -> Optional.empty());
+
+   final ExecutionGraph eg = new 
ExecutionGraphTestUtils.TestingExecutionGraphBuilder(jg)
+   .setRestartStrategy(manuallyTriggeredRestartStrategy)
+   
.setFailoverStrategyFactory(TestAdaptedRestartPipelinedRegionStrategyNG::new)
+   .setSlotProvider(slotProvider)
+   .setJobMasterConfig(jmConfig)
+   .setPartitionTracker(partitionTracker)
+   .build();
 
-   graph.start(componentMainThreadExecutor);
+   eg.start(componentMainThreadExecutor);
 
-   return graph;
+   return eg;
 
 Review comment:
   Ok.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] Leverage JM tracked partitions to find out unavailable partitions in …

2019-07-21 Thread GitBox
zhuzhurk commented on a change in pull request #9165: [FLINK-13055] [runtime] 
Leverage JM tracked partitions to find out unavailable partitions in …
URL: https://github.com/apache/flink/pull/9165#discussion_r305658771
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGConcurrentFailoverTest.java
 ##
 @@ -250,37 +253,38 @@ public void 
testSkipFailoverIfExecutionStateIsNotRunning() throws Exception {
 */
private ExecutionGraph createExecutionGraph() throws Exception {
 
-   final JobInformation jobInformation = new 
DummyJobInformation(TEST_JOB_ID, "test job");
-   final SimpleSlotProvider slotProvider = new 
SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM);
-
-   final Time timeout = Time.seconds(10L);
-   final ExecutionGraph graph = new ExecutionGraph(
-   jobInformation,
-   TestingUtils.defaultExecutor(),
-   TestingUtils.defaultExecutor(),
-   timeout,
-   manuallyTriggeredRestartStrategy,
-   TestAdaptedRestartPipelinedRegionStrategyNG::new,
-   slotProvider,
-   getClass().getClassLoader(),
-   VoidBlobWriter.getInstance(),
-   timeout);
-
-   JobVertex v1 = new JobVertex("vertex1");
+   final JobVertex v1 = new JobVertex("vertex1");
v1.setInvokableClass(NoOpInvokable.class);
v1.setParallelism(DEFAULT_PARALLELISM);
 
-   JobVertex v2 = new JobVertex("vertex2");
+   final JobVertex v2 = new JobVertex("vertex2");
v2.setInvokableClass(NoOpInvokable.class);
v2.setParallelism(DEFAULT_PARALLELISM);
 
v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, 
ResultPartitionType.BLOCKING);
 
-   JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, v2);
-   
graph.attachJobGraph(jg.getVerticesSortedTopologicallyFromSources());
+   final JobGraph jg = new JobGraph(TEST_JOB_ID, "testjob", v1, 
v2);
+
+   final SimpleSlotProvider slotProvider = new 
SimpleSlotProvider(TEST_JOB_ID, DEFAULT_PARALLELISM);
+
+   final Configuration jmConfig = new Configuration();
+   
jmConfig.setBoolean(JobManagerOptions.FORCE_PARTITION_RELEASE_ON_CONSUMPTION, 
false);
 
 Review comment:
   Yes.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-21 Thread GitBox
lirui-apache commented on issue #9181: [FLINK-13210][hive] Hive connector test 
should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-513618686
 
 
   @JingsongLi thanks for the review. One side note is blink planner requires a 
table source to implement `getReturnType()` method (hence the change to 
`HiveTableSource`). This method is marked as deprecated and seems the legacy 
planner doesn't require it. So I think it's better if we can avoid this 
inconsistency.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-13329) Set env config for sql jobs

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13329?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-13329.
---
   Resolution: Won't Fix
Fix Version/s: (was: 1.10.0)
   (was: 1.9.0)

Closing this issue, because we don't need to solve this with discussed with 
[~julien1987].

> Set env config for sql jobs
> ---
>
> Key: FLINK-13329
> URL: https://issues.apache.org/jira/browse/FLINK-13329
> Project: Flink
>  Issue Type: Task
>  Components: Table SQL / API
>Affects Versions: 1.9.0, 1.10.0
>Reporter: XuPingyong
>Priority: Major
>
> Now we execute streaming job through TableEnvironment, but
> StreamExecutionEnvironment can not be touched by users, so we can not set 
> checkpoint and other env configs when we execute sql jobs.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-10806) Support multiple consuming offsets when discovering a new topic

2019-07-21 Thread Jiayi Liao (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-10806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889876#comment-16889876
 ] 

Jiayi Liao commented on FLINK-10806:


[~becket_qin] Firstly it's not only for a newly found partition but also newly 
topic. Assuming that I've already have a streaming pipeline for processing the 
data which are collected from user behaviours on web browsers. Now I want to 
add a new topic(user behaviours on mobile devices) into this pipeline, but this 
topic has existed for a while and already been used in other pipelines. So I 
stop the pipeline and add the topic into the KafkaConsumer topic lists and 
restart this pipeline, which causes that the data of this topic is consumed 
from earliest offset because it's a newly found TopicAndPartition. I think It's 
reasonable that we're able to consume from the latest offset of this newly 
topic.

> Support multiple consuming offsets when discovering a new topic
> ---
>
> Key: FLINK-10806
> URL: https://issues.apache.org/jira/browse/FLINK-10806
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.6.2, 1.8.1
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Major
>
> In KafkaConsumerBase, we discover the TopicPartitions and compare them with 
> the restoredState. It's reasonable when a topic's partitions scaled. However, 
> if we add a new topic which has too much data and restore the Flink program, 
> the data of the new topic will be consumed from the start, which may not be 
> what we want.  I think this should be an option for developers.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13341) Some streamTableSink#consumeDataStream(DataStream) don't be implemented as returning the sink transformation DataStreamSink when using blink planner.

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13341:

Priority: Critical  (was: Major)

> Some streamTableSink#consumeDataStream(DataStream) don't be implemented as 
> returning the sink transformation DataStreamSink when using blink planner.
> -
>
> Key: FLINK-13341
> URL: https://issues.apache.org/jira/browse/FLINK-13341
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: chenqi
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.9.0, 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Some streamTableSink#consumeDataStream(DataStream) don't be implemented as 
> returning the sink transformation DataStreamSink when using blink planner.
> which will throw the following errors:
> {code:java}
> Exception in thread "main" org.apache.flink.table.api.TableException: The 
> StreamTableSink#consumeDataStream(DataStream) must be implemented and return 
> the sink transformation DataStreamSink. However, 
> org.apache.flink.streaming.connectors.kafka.Kafka010TableSink doesn't 
> implement this method.
>  at 
> org.apache.flink.table.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:123)
>  at 
> org.apache.flink.table.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:50)
>  at 
> org.apache.flink.table.plan.nodes.exec.ExecNode$class.translateToPlan(ExecNode.scala:60)
>  at 
> org.apache.flink.table.plan.nodes.physical.stream.StreamExecSink.translateToPlan(StreamExecSink.scala:50)
>  at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:71)
>  at 
> org.apache.flink.table.planner.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:70)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>  at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>  at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
>  at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>  at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
>  at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>  at 
> org.apache.flink.table.planner.StreamPlanner.translateToPlan(StreamPlanner.scala:70)
>  at 
> org.apache.flink.table.planner.PlannerBase.translate(PlannerBase.scala:155)
>  at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:446)
>  at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.sqlUpdate(TableEnvironmentImpl.java:355)
>  at 
> org.apache.flink.table.examples.java.StreamSQLLookupJoinExample.main(StreamSQLLookupJoinExample.java:139){code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-12567) Rework DescriptorProperties to adapt unified DDL with clause and Descriptor key value pairs

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-12567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-12567:

Fix Version/s: (was: 1.9.0)
   1.10.0

> Rework DescriptorProperties to adapt unified DDL with clause and Descriptor 
> key value pairs
> ---
>
> Key: FLINK-12567
> URL: https://issues.apache.org/jira/browse/FLINK-12567
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Affects Versions: 1.8.0
>Reporter: Danny Chan
>Assignee: Danny Chan
>Priority: Major
> Fix For: 1.10.0
>
>
> After introduce DDLs, we need to unify the k-v properties format for DDL with 
> clause and Descriptor API, this tool class also need to rework.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-12567) Rework DescriptorProperties to adapt unified DDL with clause and Descriptor key value pairs

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-12567?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-12567:

Affects Version/s: (was: 1.8.0)

> Rework DescriptorProperties to adapt unified DDL with clause and Descriptor 
> key value pairs
> ---
>
> Key: FLINK-12567
> URL: https://issues.apache.org/jira/browse/FLINK-12567
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Reporter: Danny Chan
>Assignee: Danny Chan
>Priority: Major
> Fix For: 1.10.0
>
>
> After introduce DDLs, we need to unify the k-v properties format for DDL with 
> clause and Descriptor API, this tool class also need to rework.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-12249) Type equivalence check fails for Window Aggregates

2019-07-21 Thread sunjincheng (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-12249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889874#comment-16889874
 ] 

sunjincheng commented on FLINK-12249:
-

And I like the new update about workaround patch, in this way we can solve this 
issue for user in 1.9, and create a new JIRA for the follow-up discussion. just 
like [~dawidwys] and all of you mentioned this issue is not the blocker of 1.9 
release. 

> Type equivalence check fails for Window Aggregates
> --
>
> Key: FLINK-12249
> URL: https://issues.apache.org/jira/browse/FLINK-12249
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Legacy Planner, Tests
>Affects Versions: 1.9.0
>Reporter: Dawid Wysakowicz
>Assignee: Hequn Cheng
>Priority: Critical
> Fix For: 1.9.0
>
>
> Creating Aggregate node fails in rules: {{LogicalWindowAggregateRule}} and 
> {{ExtendedAggregateExtractProjectRule}} if the only grouping expression is a 
> window and
> we compute aggregation on NON NULLABLE field.
> The root cause for that, is how return type inference strategies in calcite 
> work and how we handle window aggregates. Take 
> {{org.apache.calcite.sql.type.ReturnTypes#AGG_SUM}} as an example, based on 
> {{groupCount}} it adjusts type nullability based on groupCount.
> Though we pass a false information as we strip down window aggregation from 
> groupSet (in {{LogicalWindowAggregateRule}}).
> One can reproduce this problem also with a unit test like this:
> {code}
> @Test
>   def testTumbleFunction2() = {
>  
> val innerQuery =
>   """
> |SELECT
> | CASE a WHEN 1 THEN 1 ELSE 99 END AS correct,
> | rowtime
> |FROM MyTable
>   """.stripMargin
> val sql =
>   "SELECT " +
> "  SUM(correct) as cnt, " +
> "  TUMBLE_START(rowtime, INTERVAL '15' MINUTE) as wStart " +
> s"FROM ($innerQuery) " +
> "GROUP BY TUMBLE(rowtime, INTERVAL '15' MINUTE)"
> val expected = ""
> streamUtil.verifySql(sql, expected)
>   }
> {code}
> This causes e2e tests to fail: 
> https://travis-ci.org/apache/flink/builds/521183361?utm_source=slack_medium=notificationhttps://travis-ci.org/apache/flink/builds/521183361?utm_source=slack_medium=notification



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-12557) Unify create table DDL with clause and connector descriptor keys

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-12557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-12557:

Fix Version/s: (was: 1.9.0)
   1.10.0

> Unify create table DDL with clause and connector descriptor keys
> 
>
> Key: FLINK-12557
> URL: https://issues.apache.org/jira/browse/FLINK-12557
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Affects Versions: 1.8.0
>Reporter: Danny Chan
>Assignee: Danny Chan
>Priority: Major
> Fix For: 1.10.0
>
>
> The *with* option in table DDL defines the properties needed for specific 
> connector to create TableSource/Sink. The properties structure for SqlClient 
> config YAML is defined in [Improvements to the Unified SQL Connector 
> API|https://docs.google.com/document/d/1Yaxp1UJUFW-peGLt8EIidwKIZEWrrA-pznWLuvaH39Y/edit#heading=h.41fd6rs7b3cf],
>  in this design, the properties can be categorized into 4 parts:
>  
>  # Top level properties: name, type(source, sink), update-mode ...
>  # Connector specific properties: connector.type, connector.path ...
>  # Format properties: format.type, format.fields.1.name ...
>  # Table schema properties: (can be omitted for DDL)
>  
> This properties structure is reasonable for YAML, but they are not that 
> concise enough for developers.  So there also defines a tool class named 
> [DescriptorProperties|https://github.com/apache/flink/blob/b3604f7bee7456b8533e9ea222a833a2624e36c2/flink-table/flink-table-common/src/main/java/org/apache/flink/table/descriptors/DescriptorProperties.java#L67]
>  to reconstruct the data structure(like TableSchema) from the flat k-v 
> strings.
>  
> So in order to reduce complexity and keep the KV consistency for DDL with 
> properties and TableFactory properties, i proposed to simplify the DDL with 
> properties keys as following (corresponding to above 4 categories):
>  
>  # Top level properties: keep same as that in the YAML e.g. connector, 
> update-mode
>  # Connector specific properties: start with prefix named the connector type 
> e.g. for kafka connector, the properties are defined as kafka.k1 = v1, 
> kafka.k2 = v2
>  # Format properties: format.type simplified to format and the others with 
> prefix of the format name e.g. format = 'json', json.line-delimiter = "\n"
>  # Table schema properties: omitted.
> Here is a demo of creat table DDL:
> {code:java}
> CREATE TABLE Kafka10SourceTable (
>   intField INTEGER,
>   stringField VARCHAR(128) COMMENT 'User IP address',
>   longField BIGINT,
>   rowTimeField TIMESTAMP,
>   WATERMARK wm01 FOR  'longField' AS BOUNDED WITH DELAY '60' SECOND
> )
> COMMENT 'Kafka Source Table of topic user_ip_address'
> WITH (
>   connector='kafka',
>   kafka.property-version='1',
>   kafka.version='0.10',
>   kafka.topic='test-kafka-topic',
>   kafka.startup-mode = 'latest-offset'
>   kafka.specific-offset = 'offset'
>   format='json'
>   json.property-version = '1'
>   json.version='1'
>   json.derive-schema='true'
> )
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13105) Add documentation for blink planner's built-in functions

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13105:

Priority: Critical  (was: Major)

> Add documentation for blink planner's built-in functions
> 
>
> Key: FLINK-13105
> URL: https://issues.apache.org/jira/browse/FLINK-13105
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Reporter: Zhenghua Gao
>Assignee: Zhenghua Gao
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.9.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Blink planner intros some built-in functions which need to be documented.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-12557) Unify create table DDL with clause and connector descriptor keys

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-12557?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-12557:

Affects Version/s: (was: 1.8.0)

> Unify create table DDL with clause and connector descriptor keys
> 
>
> Key: FLINK-12557
> URL: https://issues.apache.org/jira/browse/FLINK-12557
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API
>Reporter: Danny Chan
>Assignee: Danny Chan
>Priority: Major
> Fix For: 1.10.0
>
>
> The *with* option in table DDL defines the properties needed for specific 
> connector to create TableSource/Sink. The properties structure for SqlClient 
> config YAML is defined in [Improvements to the Unified SQL Connector 
> API|https://docs.google.com/document/d/1Yaxp1UJUFW-peGLt8EIidwKIZEWrrA-pznWLuvaH39Y/edit#heading=h.41fd6rs7b3cf],
>  in this design, the properties can be categorized into 4 parts:
>  
>  # Top level properties: name, type(source, sink), update-mode ...
>  # Connector specific properties: connector.type, connector.path ...
>  # Format properties: format.type, format.fields.1.name ...
>  # Table schema properties: (can be omitted for DDL)
>  
> This properties structure is reasonable for YAML, but they are not that 
> concise enough for developers.  So there also defines a tool class named 
> [DescriptorProperties|https://github.com/apache/flink/blob/b3604f7bee7456b8533e9ea222a833a2624e36c2/flink-table/flink-table-common/src/main/java/org/apache/flink/table/descriptors/DescriptorProperties.java#L67]
>  to reconstruct the data structure(like TableSchema) from the flat k-v 
> strings.
>  
> So in order to reduce complexity and keep the KV consistency for DDL with 
> properties and TableFactory properties, i proposed to simplify the DDL with 
> properties keys as following (corresponding to above 4 categories):
>  
>  # Top level properties: keep same as that in the YAML e.g. connector, 
> update-mode
>  # Connector specific properties: start with prefix named the connector type 
> e.g. for kafka connector, the properties are defined as kafka.k1 = v1, 
> kafka.k2 = v2
>  # Format properties: format.type simplified to format and the others with 
> prefix of the format name e.g. format = 'json', json.line-delimiter = "\n"
>  # Table schema properties: omitted.
> Here is a demo of creat table DDL:
> {code:java}
> CREATE TABLE Kafka10SourceTable (
>   intField INTEGER,
>   stringField VARCHAR(128) COMMENT 'User IP address',
>   longField BIGINT,
>   rowTimeField TIMESTAMP,
>   WATERMARK wm01 FOR  'longField' AS BOUNDED WITH DELAY '60' SECOND
> )
> COMMENT 'Kafka Source Table of topic user_ip_address'
> WITH (
>   connector='kafka',
>   kafka.property-version='1',
>   kafka.version='0.10',
>   kafka.topic='test-kafka-topic',
>   kafka.startup-mode = 'latest-offset'
>   kafka.specific-offset = 'offset'
>   format='json'
>   json.property-version = '1'
>   json.version='1'
>   json.derive-schema='true'
> )
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13300) Add OverwritableTableSink bridge logic to flink planner

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13300:

Fix Version/s: (was: 1.9.0)

> Add OverwritableTableSink bridge logic to flink planner
> -
>
> Key: FLINK-13300
> URL: https://issues.apache.org/jira/browse/FLINK-13300
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.9.0
>Reporter: Danny Chan
>Assignee: Danny Chan
>Priority: Major
> Fix For: 1.10.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13300) Add OverwritableTableSink bridge logic to flink planner

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13300?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-13300:

Affects Version/s: (was: 1.9.0)

> Add OverwritableTableSink bridge logic to flink planner
> -
>
> Key: FLINK-13300
> URL: https://issues.apache.org/jira/browse/FLINK-13300
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Danny Chan
>Assignee: Danny Chan
>Priority: Major
> Fix For: 1.10.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] sunjincheng121 commented on issue #9141: [FLINK-12249][table] Fix type equivalence check problems for Window Aggregates

2019-07-21 Thread GitBox
sunjincheng121 commented on issue #9141: [FLINK-12249][table] Fix type 
equivalence check problems for Window Aggregates
URL: https://github.com/apache/flink/pull/9141#issuecomment-513617339
 
 
   Thanks for the update @hequn8128!
   I agree with @dawidwys @hequn8128  that this issue is not the blocker of 1.9 
release.
   
   And I like this workaround patch, in this way we can solve this issue for 
user in 1.9, and create a new JIRA for  the follow-up discussion.
   
   What do you think?
   
   Best, Jincheng
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13177) make FunctionCatalog able to reference functions across catalogs

2019-07-21 Thread Jark Wu (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889870#comment-16889870
 ] 

Jark Wu commented on FLINK-13177:
-

Hi [~phoenixjiangnan], do we need this in 1.9.0? 

> make FunctionCatalog able to reference functions across catalogs
> 
>
> Key: FLINK-13177
> URL: https://issues.apache.org/jira/browse/FLINK-13177
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Reporter: Bowen Li
>Assignee: Bowen Li
>Priority: Major
> Fix For: 1.10.0
>
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-11433) JOIN on a table having a column of type MULTISET gives a NPE

2019-07-21 Thread Jark Wu (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-11433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-11433:

Priority: Minor  (was: Major)

>  JOIN on a table having a column of type MULTISET gives a NPE
> -
>
> Key: FLINK-11433
> URL: https://issues.apache.org/jira/browse/FLINK-11433
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.7.0, 1.7.1
>Reporter: Elias Saalmann
>Assignee: TANG Wen-hui
>Priority: Minor
>
> I get an error (Error while applying rule FlinkLogicalJoinConverter) when 
> performing a JOIN on a table having a column of type MULTISET (e.g. a COLLECT 
> as aggregation of a GROUP BY), for instance:
> SELECT a, d
>  FROM TableA JOIN (
>    SELECT b, COLLECT(c) AS d
>    FROM TableB
>    GROUP BY b
>  ) TableC ON a = b
> Full stacktrace:
> Exception in thread "main" java.lang.RuntimeException: Error while applying 
> rule FlinkLogicalJoinConverter, args 
> [rel#71:LogicalJoin.NONE(left=rel#69:Subset#3.NONE,right=rel#70:Subset#4.NONE,condition==($2,
>  $0),joinType=inner)]
>      at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:236)
>      at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:646)
>      at 
> org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:339)
>      at 
> org.apache.flink.table.api.TableEnvironment.runVolcanoPlanner(TableEnvironment.scala:373)
>      at 
> org.apache.flink.table.api.TableEnvironment.optimizeLogicalPlan(TableEnvironment.scala:292)
>      at 
> org.apache.flink.table.api.BatchTableEnvironment.optimize(BatchTableEnvironment.scala:455)
>      at 
> org.apache.flink.table.api.BatchTableEnvironment.translate(BatchTableEnvironment.scala:475)
>      at 
> org.apache.flink.table.api.java.BatchTableEnvironment.toDataSet(BatchTableEnvironment.scala:165)
>      at org.myorg.quickstart.TableJob2.main(TableJob2.java:40)
>  Caused by: java.lang.RuntimeException: Error occurred while applying rule 
> FlinkLogicalJoinConverter
>      at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.transformTo(VolcanoRuleCall.java:149)
>      at 
> org.apache.calcite.plan.RelOptRuleCall.transformTo(RelOptRuleCall.java:234)
>      at 
> org.apache.calcite.rel.convert.ConverterRule.onMatch(ConverterRule.java:141)
>      at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:212)
>      ... 8 more
>  Caused by: java.lang.NullPointerException
>      at 
> org.apache.flink.table.plan.nodes.FlinkRelNode$class.estimateDataTypeSize(FlinkRelNode.scala:84)
>      at 
> org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.estimateDataTypeSize(FlinkLogicalJoinBase.scala:29)
>      at 
> org.apache.flink.table.plan.nodes.FlinkRelNode$class.estimateDataTypeSize(FlinkRelNode.scala:104)
>      at 
> org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.estimateDataTypeSize(FlinkLogicalJoinBase.scala:29)
>      at 
> org.apache.flink.table.plan.nodes.FlinkRelNode$$anonfun$estimateRowSize$2.apply(FlinkRelNode.scala:80)
>      at 
> org.apache.flink.table.plan.nodes.FlinkRelNode$$anonfun$estimateRowSize$2.apply(FlinkRelNode.scala:79)
>      at 
> scala.collection.IndexedSeqOptimized$class.foldl(IndexedSeqOptimized.scala:57)
>      at 
> scala.collection.IndexedSeqOptimized$class.foldLeft(IndexedSeqOptimized.scala:66)
>      at scala.collection.mutable.ArrayBuffer.foldLeft(ArrayBuffer.scala:48)
>      at 
> org.apache.flink.table.plan.nodes.FlinkRelNode$class.estimateRowSize(FlinkRelNode.scala:79)
>      at 
> org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.estimateRowSize(FlinkLogicalJoinBase.scala:29)
>      at 
> org.apache.flink.table.plan.nodes.logical.FlinkLogicalJoinBase.computeSelfCost(FlinkLogicalJoinBase.scala:48)
>      at 
> org.apache.calcite.rel.metadata.RelMdPercentageOriginalRows.getNonCumulativeCost(RelMdPercentageOriginalRows.java:162)
>      at 
> GeneratedMetadataHandler_NonCumulativeCost.getNonCumulativeCost_$(Unknown 
> Source)
>      at 
> GeneratedMetadataHandler_NonCumulativeCost.getNonCumulativeCost(Unknown 
> Source)
>      at 
> org.apache.calcite.rel.metadata.RelMetadataQuery.getNonCumulativeCost(RelMetadataQuery.java:301)
>      at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.getCost(VolcanoPlanner.java:953)
>      at 
> org.apache.calcite.plan.volcano.RelSubset.propagateCostImprovements0(RelSubset.java:339)
>      at 
> org.apache.calcite.plan.volcano.RelSubset.propagateCostImprovements(RelSubset.java:322)
>      at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.addRelToSet(VolcanoPlanner.java:1643)
>      at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.registerImpl(VolcanoPlanner.java:1579)
>      at 
> 

[GitHub] [flink] JingsongLi commented on issue #9181: [FLINK-13210][hive] Hive connector test should dependent on blink planner instead of legacy planner

2019-07-21 Thread GitBox
JingsongLi commented on issue #9181: [FLINK-13210][hive] Hive connector test 
should dependent on blink planner instead of legacy planner
URL: https://github.com/apache/flink/pull/9181#issuecomment-513613380
 
 
   @lirui-apache Thanks for your contribution, LGTM


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] Add CachedMiniCluster to share cluster between ITCases

2019-07-21 Thread GitBox
flinkbot edited a comment on issue #9180: [FLINK-13331][table-planner-blink] 
Add CachedMiniCluster to share cluster between ITCases
URL: https://github.com/apache/flink/pull/9180#issuecomment-513163961
 
 
   ## CI report:
   
   * f563446dc5923ef86883f03f5fe899c4ffc92c57 : FAILURE 
[Build](https://travis-ci.com/flink-ci/flink/builds/119766053)
   * 228fb660ad73180cdbf3abee5deeeb58b52b4d12 : PENDING 
[Build](https://travis-ci.com/flink-ci/flink/builds/119948346)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (FLINK-13231) Add a ratelimiter to pubsub source

2019-07-21 Thread Jiangjie Qin (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889868#comment-16889868
 ] 

Jiangjie Qin edited comment on FLINK-13231 at 7/22/19 1:54 AM:
---

[~Xeli] Sorry for the late reply. The rate limiter is actually a per consumer / 
per sub-task rate limiter. I guess the Java doc is a little confusing, but that 
is actually what it is.


was (Author: becket_qin):
[~Xeli] The rate limiter is actually a per consumer / per sub-task rate 
limiter. I guess the Java doc is a little confusing, but that is actually what 
it is.

> Add a ratelimiter to pubsub source
> --
>
> Key: FLINK-13231
> URL: https://issues.apache.org/jira/browse/FLINK-13231
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Google Cloud PubSub
>Reporter: Richard Deurwaarder
>Priority: Minor
>
> Replace MaxMessagesToAcknowledge limit by introducing a rate limiter. See: 
> [https://github.com/apache/flink/pull/6594#discussion_r300215868]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13231) Add a ratelimiter to pubsub source

2019-07-21 Thread Jiangjie Qin (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889868#comment-16889868
 ] 

Jiangjie Qin commented on FLINK-13231:
--

[~Xeli] The rate limiter is actually a per consumer / per sub-task rate 
limiter. I guess the Java doc is a little confusing, but that is actually what 
it is.

> Add a ratelimiter to pubsub source
> --
>
> Key: FLINK-13231
> URL: https://issues.apache.org/jira/browse/FLINK-13231
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Google Cloud PubSub
>Reporter: Richard Deurwaarder
>Priority: Minor
>
> Replace MaxMessagesToAcknowledge limit by introducing a rate limiter. See: 
> [https://github.com/apache/flink/pull/6594#discussion_r300215868]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305650403
 
 

 ##
 File path: 
flink-streaming-java/src/test/java/org/apache/flink/streaming/api/graph/StreamingJobGraphGeneratorTest.java
 ##
 @@ -593,12 +593,19 @@ public void testDefaultScheduleMode() {
public void testSetScheduleMode() {
StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment();
 
-   StreamGraph streamGraph = new 
StreamGraphGenerator(Collections.emptyList(),
+   StreamGraph streamGraph1 = new 
StreamGraphGenerator(Collections.emptyList(),
env.getConfig(), env.getCheckpointConfig())
.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES)
.generate();
-   JobGraph jobGraph = 
StreamingJobGraphGenerator.createJobGraph(streamGraph);
-   assertEquals(ScheduleMode.LAZY_FROM_SOURCES, 
jobGraph.getScheduleMode());
+   JobGraph jobGraph1 = 
StreamingJobGraphGenerator.createJobGraph(streamGraph1);
+   assertEquals(ScheduleMode.LAZY_FROM_SOURCES, 
jobGraph1.getScheduleMode());
+
+   StreamGraph streamGraph2 = new 
StreamGraphGenerator(Collections.emptyList(),
+   env.getConfig(), env.getCheckpointConfig())
+   
.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST)
 
 Review comment:
   unnecessary changes?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305650383
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/minicluster/MiniClusterITCase.java
 ##
 @@ -117,7 +117,7 @@ public void testHandleStreamingJobsWhenNotEnoughSlot() 
throws Exception {
@Test
public void testHandleBatchJobsWhenNotEnoughSlot() throws Exception {
try {
-   
setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES);
+   
setupAndRunHandleJobsWhenNotEnoughSlots(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST);
 
 Review comment:
   unnecessary changes?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305650367
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/jobgraph/ScheduleModeTest.java
 ##
 @@ -31,6 +31,7 @@
@Test
public void testAllowLazyDeployment() throws Exception {

assertTrue(ScheduleMode.LAZY_FROM_SOURCES.allowLazyDeployment());
+   
assertTrue(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST.allowLazyDeployment());
 
 Review comment:
   unnecessary changes?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL / Planner] Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch

2019-07-21 Thread GitBox
KurtYoung commented on a change in pull request #9101: [FLINK-13221][Table SQL 
/ Planner] Blink planner should set ScheduleMode to 
LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
URL: https://github.com/apache/flink/pull/9101#discussion_r305650341
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/AdaptedRestartPipelinedRegionStrategyNGFailoverTest.java
 ##
 @@ -366,7 +366,7 @@ private JobGraph createBatchJobGraph() {
v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL, 
ResultPartitionType.BLOCKING);
 
final JobGraph jobGraph = new JobGraph(v1, v2);
-   jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES);
+   
jobGraph.setScheduleMode(ScheduleMode.LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST);
 
 Review comment:
   unnecessary changes?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread Kurt Young (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889864#comment-16889864
 ] 

Kurt Young commented on FLINK-13221:


I changed this to 1.9.0's blocker, since this is a crucial one to fully turn on 
runtime's fine grained recovery for batch jobs.

[~fan_li_ya] Could you also open a PR to 1.9 branch?

> Blink planner should set ScheduleMode to 
> LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
> -
>
> Key: FLINK-13221
> URL: https://issues.apache.org/jira/browse/FLINK-13221
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Kurt Young
>Assignee: Liya Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0, 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread Kurt Young (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-13221:
---
Fix Version/s: 1.10.0
   1.9.0

> Blink planner should set ScheduleMode to 
> LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
> -
>
> Key: FLINK-13221
> URL: https://issues.apache.org/jira/browse/FLINK-13221
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Kurt Young
>Assignee: Liya Fan
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.9.0, 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread Kurt Young (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-13221:
---
Priority: Blocker  (was: Major)

> Blink planner should set ScheduleMode to 
> LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
> -
>
> Key: FLINK-13221
> URL: https://issues.apache.org/jira/browse/FLINK-13221
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Kurt Young
>Assignee: Liya Fan
>Priority: Blocker
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Updated] (FLINK-13221) Blink planner should set ScheduleMode to LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs

2019-07-21 Thread Kurt Young (JIRA)


 [ 
https://issues.apache.org/jira/browse/FLINK-13221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-13221:
---
Issue Type: Bug  (was: Improvement)

> Blink planner should set ScheduleMode to 
> LAZY_FROM_SOURCES_WITH_BATCH_SLOT_REQUEST for batch jobs
> -
>
> Key: FLINK-13221
> URL: https://issues.apache.org/jira/browse/FLINK-13221
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Kurt Young
>Assignee: Liya Fan
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[jira] [Commented] (FLINK-13199) ARM support for Flink

2019-07-21 Thread wangxiyuan (JIRA)


[ 
https://issues.apache.org/jira/browse/FLINK-13199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16889861#comment-16889861
 ] 

wangxiyuan commented on FLINK-13199:


Hi Chesnay:

 

It's similar. The difference is that which CI system will be used.

OpenLab has two ways to support ARM build for Apache community.

One is that Openlab provides CI system directly, like travis-ci, Flink can add 
`openlab-ci`[1] as the ARM build CI the same as travis-ci for X86.

The other ways is what is said on Infra ML, Openlab can donate ARM resources to 
Infra team, and then Flink can use the apache offical CI system[2] as the ARM 
build CI.

 

For the first way, OpenLab team can help to write CI jobs scripts and keep 
maintaining the related CI jobs.

For the second way, I guess Infra team will handle it.

[1]: [https://status.openlabtesting.org/status]

[2]: [https://builds.apache.org/]

> ARM support for Flink
> -
>
> Key: FLINK-13199
> URL: https://issues.apache.org/jira/browse/FLINK-13199
> Project: Flink
>  Issue Type: New Feature
>  Components: Build System
>Reporter: wangxiyuan
>Priority: Critical
>
> There is not official ARM release for Flink. But basing on my local test, 
> Flink which is made by Java and Scala is built and tested well. So is it 
> possible to support ARM release officially? And I think it's may not be a 
> huge work.
>  
> AFAIK, Flink now uses travis-ci which supports only x86 for CI gate. Is it 
> possible to add an ARM one? I'm from openlab community[1]. Similar with 
> travis-ci, it's is an opensource and free community which provide CI 
> resources and system for opensource projects, contains both ARM and X86 
> machines. And now it helps some community building there CI already. Such as 
> OpenStack and CNCF.
>  
> If Flink community agree to support ARM. I can spend my full time to help. 
> Such as job define, CI maintaining, test fix and so on. If Flink don't want 
> to rely on OpenLab, we can donate ARM resources directly as well.
>  
> I have sent out a discuess mail-list already[2]. Feel free to reply there or 
> here.
>  
> Thanks.
>  
> [1]:[https://openlabtesting.org/]
> [2]:[http://mail-archives.apache.org/mod_mbox/flink-dev/201907.mbox/browser]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)


[GitHub] [flink] TsReaper commented on a change in pull request #9139: [FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation of getString and getBinary method in NestedRow, f

2019-07-21 Thread GitBox
TsReaper commented on a change in pull request #9139: 
[FLINK-13304][FLINK-13322][FLINK-13323][table-runtime-blink] Fix implementation 
of getString and getBinary method in NestedRow, fix serializer restore in 
BaseArray/Map serializer and add tests for complex data formats
URL: https://github.com/apache/flink/pull/9139#discussion_r305649009
 
 

 ##
 File path: 
flink-core/src/test/java/org/apache/flink/api/common/typeutils/SerializerTestBase.java
 ##
 @@ -124,30 +124,8 @@ public void testSnapshotConfigurationAndReconfigure() 
throws Exception {
final TypeSerializer serializer = getSerializer();
final TypeSerializerSnapshot configSnapshot = 
serializer.snapshotConfiguration();
 
-   byte[] serializedConfig;
-   try (ByteArrayOutputStream out = new ByteArrayOutputStream()) {
-   
TypeSerializerSnapshotSerializationUtil.writeSerializerSnapshot(
-   new DataOutputViewStreamWrapper(out), 
configSnapshot, serializer);
-   serializedConfig = out.toByteArray();
-   }
-
-   TypeSerializerSnapshot restoredConfig;
-   try (ByteArrayInputStream in = new 
ByteArrayInputStream(serializedConfig)) {
-   restoredConfig = 
TypeSerializerSnapshotSerializationUtil.readSerializerSnapshot(
-   new DataInputViewStreamWrapper(in), 
Thread.currentThread().getContextClassLoader(), getSerializer());
-   }
-
-   TypeSerializerSchemaCompatibility strategy = 
restoredConfig.resolveSchemaCompatibility(getSerializer());
-   final TypeSerializer restoreSerializer;
-   if (strategy.isCompatibleAsIs()) {
-   restoreSerializer = restoredConfig.restoreSerializer();
-   }
-   else if (strategy.isCompatibleWithReconfiguredSerializer()) {
-   restoreSerializer = 
strategy.getReconfiguredSerializer();
-   }
-   else {
-   throw new AssertionError("Unable to restore serializer 
with " + strategy);
-   }
+   final TypeSerializer restoreSerializer = 
snapshotAndReconfigure(
 
 Review comment:
   No. I just move these into a separate method so that it can be reused for my 
newly added tests.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] sjwiesman edited a comment on issue #9161: [FLINK-13262][docs] Add documentation for the new Table & SQL API type system

2019-07-21 Thread GitBox
sjwiesman edited a comment on issue #9161: [FLINK-13262][docs] Add 
documentation for the new Table & SQL API type system
URL: https://github.com/apache/flink/pull/9161#issuecomment-513228516
 
 
   @twalthr Looks good to me +1


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] tweise commented on issue #9187: [FLINK-12595][kinesis] Interrupt thread at right time to avoid deadlock

2019-07-21 Thread GitBox
tweise commented on issue #9187: [FLINK-12595][kinesis] Interrupt thread at 
right time to avoid deadlock
URL: https://github.com/apache/flink/pull/9187#issuecomment-513580704
 
 
   @flinkbot approve all


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


  1   2   >